linux-yocto/tools/testing/selftests/mm/mseal_test.c
Linus Torvalds 617a814f14 ALong with the usual shower of singleton patches, notable patch series in
this pull request are:
 
 "Align kvrealloc() with krealloc()" from Danilo Krummrich.  Adds
 consistency to the APIs and behaviour of these two core allocation
 functions.  This also simplifies/enables Rustification.
 
 "Some cleanups for shmem" from Baolin Wang.  No functional changes - mode
 code reuse, better function naming, logic simplifications.
 
 "mm: some small page fault cleanups" from Josef Bacik.  No functional
 changes - code cleanups only.
 
 "Various memory tiering fixes" from Zi Yan.  A small fix and a little
 cleanup.
 
 "mm/swap: remove boilerplate" from Yu Zhao.  Code cleanups and
 simplifications and .text shrinkage.
 
 "Kernel stack usage histogram" from Pasha Tatashin and Shakeel Butt.  This
 is a feature, it adds new feilds to /proc/vmstat such as
 
     $ grep kstack /proc/vmstat
     kstack_1k 3
     kstack_2k 188
     kstack_4k 11391
     kstack_8k 243
     kstack_16k 0
 
 which tells us that 11391 processes used 4k of stack while none at all
 used 16k.  Useful for some system tuning things, but partivularly useful
 for "the dynamic kernel stack project".
 
 "kmemleak: support for percpu memory leak detect" from Pavel Tikhomirov.
 Teaches kmemleak to detect leaksage of percpu memory.
 
 "mm: memcg: page counters optimizations" from Roman Gushchin.  "3
 independent small optimizations of page counters".
 
 "mm: split PTE/PMD PT table Kconfig cleanups+clarifications" from David
 Hildenbrand.  Improves PTE/PMD splitlock detection, makes powerpc/8xx work
 correctly by design rather than by accident.
 
 "mm: remove arch_make_page_accessible()" from David Hildenbrand.  Some
 folio conversions which make arch_make_page_accessible() unneeded.
 
 "mm, memcg: cg2 memory{.swap,}.peak write handlers" fro David Finkel.
 Cleans up and fixes our handling of the resetting of the cgroup/process
 peak-memory-use detector.
 
 "Make core VMA operations internal and testable" from Lorenzo Stoakes.
 Rationalizaion and encapsulation of the VMA manipulation APIs.  With a
 view to better enable testing of the VMA functions, even from a
 userspace-only harness.
 
 "mm: zswap: fixes for global shrinker" from Takero Funaki.  Fix issues in
 the zswap global shrinker, resulting in improved performance.
 
 "mm: print the promo watermark in zoneinfo" from Kaiyang Zhao.  Fill in
 some missing info in /proc/zoneinfo.
 
 "mm: replace follow_page() by folio_walk" from David Hildenbrand.  Code
 cleanups and rationalizations (conversion to folio_walk()) resulting in
 the removal of follow_page().
 
 "improving dynamic zswap shrinker protection scheme" from Nhat Pham.  Some
 tuning to improve zswap's dynamic shrinker.  Significant reductions in
 swapin and improvements in performance are shown.
 
 "mm: Fix several issues with unaccepted memory" from Kirill Shutemov.
 Improvements to the new unaccepted memory feature,
 
 "mm/mprotect: Fix dax puds" from Peter Xu.  Implements mprotect on DAX
 PUDs.  This was missing, although nobody seems to have notied yet.
 
 "Introduce a store type enum for the Maple tree" from Sidhartha Kumar.
 Cleanups and modest performance improvements for the maple tree library
 code.
 
 "memcg: further decouple v1 code from v2" from Shakeel Butt.  Move more
 cgroup v1 remnants away from the v2 memcg code.
 
 "memcg: initiate deprecation of v1 features" from Shakeel Butt.  Adds
 various warnings telling users that memcg v1 features are deprecated.
 
 "mm: swap: mTHP swap allocator base on swap cluster order" from Chris Li.
 Greatly improves the success rate of the mTHP swap allocation.
 
 "mm: introduce numa_memblks" from Mike Rapoport.  Moves various disparate
 per-arch implementations of numa_memblk code into generic code.
 
 "mm: batch free swaps for zap_pte_range()" from Barry Song.  Greatly
 improves the performance of munmap() of swap-filled ptes.
 
 "support large folio swap-out and swap-in for shmem" from Baolin Wang.
 With this series we no longer split shmem large folios into simgle-page
 folios when swapping out shmem.
 
 "mm/hugetlb: alloc/free gigantic folios" from Yu Zhao.  Nice performance
 improvements and code reductions for gigantic folios.
 
 "support shmem mTHP collapse" from Baolin Wang.  Adds support for
 khugepaged's collapsing of shmem mTHP folios.
 
 "mm: Optimize mseal checks" from Pedro Falcato.  Fixes an mprotect()
 performance regression due to the addition of mseal().
 
 "Increase the number of bits available in page_type" from Matthew Wilcox.
 Increases the number of bits available in page_type!
 
 "Simplify the page flags a little" from Matthew Wilcox.  Many legacy page
 flags are now folio flags, so the page-based flags and their
 accessors/mutators can be removed.
 
 "mm: store zero pages to be swapped out in a bitmap" from Usama Arif.  An
 optimization which permits us to avoid writing/reading zero-filled zswap
 pages to backing store.
 
 "Avoid MAP_FIXED gap exposure" from Liam Howlett.  Fixes a race window
 which occurs when a MAP_FIXED operqtion is occurring during an unrelated
 vma tree walk.
 
 "mm: remove vma_merge()" from Lorenzo Stoakes.  Major rotorooting of the
 vma_merge() functionality, making ot cleaner, more testable and better
 tested.
 
 "misc fixups for DAMON {self,kunit} tests" from SeongJae Park.  Minor
 fixups of DAMON selftests and kunit tests.
 
 "mm: memory_hotplug: improve do_migrate_range()" from Kefeng Wang.  Code
 cleanups and folio conversions.
 
 "Shmem mTHP controls and stats improvements" from Ryan Roberts.  Cleanups
 for shmem controls and stats.
 
 "mm: count the number of anonymous THPs per size" from Barry Song.  Expose
 additional anon THP stats to userspace for improved tuning.
 
 "mm: finish isolate/putback_lru_page()" from Kefeng Wang: more folio
 conversions and removal of now-unused page-based APIs.
 
 "replace per-quota region priorities histogram buffer with per-context
 one" from SeongJae Park.  DAMON histogram rationalization.
 
 "Docs/damon: update GitHub repo URLs and maintainer-profile" from SeongJae
 Park.  DAMON documentation updates.
 
 "mm/vdpa: correct misuse of non-direct-reclaim __GFP_NOFAIL and improve
 related doc and warn" from Jason Wang: fixes usage of page allocator
 __GFP_NOFAIL and GFP_ATOMIC flags.
 
 "mm: split underused THPs" from Yu Zhao.  Improve THP=always policy - this
 was overprovisioning THPs in sparsely accessed memory areas.
 
 "zram: introduce custom comp backends API" frm Sergey Senozhatsky.  Add
 support for zram run-time compression algorithm tuning.
 
 "mm: Care about shadow stack guard gap when getting an unmapped area" from
 Mark Brown.  Fix up the various arch_get_unmapped_area() implementations
 to better respect guard areas.
 
 "Improve mem_cgroup_iter()" from Kinsey Ho.  Improve the reliability of
 mem_cgroup_iter() and various code cleanups.
 
 "mm: Support huge pfnmaps" from Peter Xu.  Extends the usage of huge
 pfnmap support.
 
 "resource: Fix region_intersects() vs add_memory_driver_managed()" from
 Huang Ying.  Fix a bug in region_intersects() for systems with CXL memory.
 
 "mm: hwpoison: two more poison recovery" from Kefeng Wang.  Teaches a
 couple more code paths to correctly recover from the encountering of
 poisoned memry.
 
 "mm: enable large folios swap-in support" from Barry Song.  Support the
 swapin of mTHP memory into appropriately-sized folios, rather than into
 single-page folios.
 -----BEGIN PGP SIGNATURE-----
 
 iHUEABYIAB0WIQTTMBEPP41GrTpTJgfdBJ7gKXxAjgUCZu1BBwAKCRDdBJ7gKXxA
 jlWNAQDYlqQLun7bgsAN4sSvi27VUuWv1q70jlMXTfmjJAvQqwD/fBFVR6IOOiw7
 AkDbKWP2k0hWPiNJBGwoqxdHHx09Xgo=
 =s0T+
 -----END PGP SIGNATURE-----

Merge tag 'mm-stable-2024-09-20-02-31' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm

Pull MM updates from Andrew Morton:
 "Along with the usual shower of singleton patches, notable patch series
  in this pull request are:

   - "Align kvrealloc() with krealloc()" from Danilo Krummrich. Adds
     consistency to the APIs and behaviour of these two core allocation
     functions. This also simplifies/enables Rustification.

   - "Some cleanups for shmem" from Baolin Wang. No functional changes -
     mode code reuse, better function naming, logic simplifications.

   - "mm: some small page fault cleanups" from Josef Bacik. No
     functional changes - code cleanups only.

   - "Various memory tiering fixes" from Zi Yan. A small fix and a
     little cleanup.

   - "mm/swap: remove boilerplate" from Yu Zhao. Code cleanups and
     simplifications and .text shrinkage.

   - "Kernel stack usage histogram" from Pasha Tatashin and Shakeel
     Butt. This is a feature, it adds new feilds to /proc/vmstat such as

       $ grep kstack /proc/vmstat
       kstack_1k 3
       kstack_2k 188
       kstack_4k 11391
       kstack_8k 243
       kstack_16k 0

     which tells us that 11391 processes used 4k of stack while none at
     all used 16k. Useful for some system tuning things, but
     partivularly useful for "the dynamic kernel stack project".

   - "kmemleak: support for percpu memory leak detect" from Pavel
     Tikhomirov. Teaches kmemleak to detect leaksage of percpu memory.

   - "mm: memcg: page counters optimizations" from Roman Gushchin. "3
     independent small optimizations of page counters".

   - "mm: split PTE/PMD PT table Kconfig cleanups+clarifications" from
     David Hildenbrand. Improves PTE/PMD splitlock detection, makes
     powerpc/8xx work correctly by design rather than by accident.

   - "mm: remove arch_make_page_accessible()" from David Hildenbrand.
     Some folio conversions which make arch_make_page_accessible()
     unneeded.

   - "mm, memcg: cg2 memory{.swap,}.peak write handlers" fro David
     Finkel. Cleans up and fixes our handling of the resetting of the
     cgroup/process peak-memory-use detector.

   - "Make core VMA operations internal and testable" from Lorenzo
     Stoakes. Rationalizaion and encapsulation of the VMA manipulation
     APIs. With a view to better enable testing of the VMA functions,
     even from a userspace-only harness.

   - "mm: zswap: fixes for global shrinker" from Takero Funaki. Fix
     issues in the zswap global shrinker, resulting in improved
     performance.

   - "mm: print the promo watermark in zoneinfo" from Kaiyang Zhao. Fill
     in some missing info in /proc/zoneinfo.

   - "mm: replace follow_page() by folio_walk" from David Hildenbrand.
     Code cleanups and rationalizations (conversion to folio_walk())
     resulting in the removal of follow_page().

   - "improving dynamic zswap shrinker protection scheme" from Nhat
     Pham. Some tuning to improve zswap's dynamic shrinker. Significant
     reductions in swapin and improvements in performance are shown.

   - "mm: Fix several issues with unaccepted memory" from Kirill
     Shutemov. Improvements to the new unaccepted memory feature,

   - "mm/mprotect: Fix dax puds" from Peter Xu. Implements mprotect on
     DAX PUDs. This was missing, although nobody seems to have notied
     yet.

   - "Introduce a store type enum for the Maple tree" from Sidhartha
     Kumar. Cleanups and modest performance improvements for the maple
     tree library code.

   - "memcg: further decouple v1 code from v2" from Shakeel Butt. Move
     more cgroup v1 remnants away from the v2 memcg code.

   - "memcg: initiate deprecation of v1 features" from Shakeel Butt.
     Adds various warnings telling users that memcg v1 features are
     deprecated.

   - "mm: swap: mTHP swap allocator base on swap cluster order" from
     Chris Li. Greatly improves the success rate of the mTHP swap
     allocation.

   - "mm: introduce numa_memblks" from Mike Rapoport. Moves various
     disparate per-arch implementations of numa_memblk code into generic
     code.

   - "mm: batch free swaps for zap_pte_range()" from Barry Song. Greatly
     improves the performance of munmap() of swap-filled ptes.

   - "support large folio swap-out and swap-in for shmem" from Baolin
     Wang. With this series we no longer split shmem large folios into
     simgle-page folios when swapping out shmem.

   - "mm/hugetlb: alloc/free gigantic folios" from Yu Zhao. Nice
     performance improvements and code reductions for gigantic folios.

   - "support shmem mTHP collapse" from Baolin Wang. Adds support for
     khugepaged's collapsing of shmem mTHP folios.

   - "mm: Optimize mseal checks" from Pedro Falcato. Fixes an mprotect()
     performance regression due to the addition of mseal().

   - "Increase the number of bits available in page_type" from Matthew
     Wilcox. Increases the number of bits available in page_type!

   - "Simplify the page flags a little" from Matthew Wilcox. Many legacy
     page flags are now folio flags, so the page-based flags and their
     accessors/mutators can be removed.

   - "mm: store zero pages to be swapped out in a bitmap" from Usama
     Arif. An optimization which permits us to avoid writing/reading
     zero-filled zswap pages to backing store.

   - "Avoid MAP_FIXED gap exposure" from Liam Howlett. Fixes a race
     window which occurs when a MAP_FIXED operqtion is occurring during
     an unrelated vma tree walk.

   - "mm: remove vma_merge()" from Lorenzo Stoakes. Major rotorooting of
     the vma_merge() functionality, making ot cleaner, more testable and
     better tested.

   - "misc fixups for DAMON {self,kunit} tests" from SeongJae Park.
     Minor fixups of DAMON selftests and kunit tests.

   - "mm: memory_hotplug: improve do_migrate_range()" from Kefeng Wang.
     Code cleanups and folio conversions.

   - "Shmem mTHP controls and stats improvements" from Ryan Roberts.
     Cleanups for shmem controls and stats.

   - "mm: count the number of anonymous THPs per size" from Barry Song.
     Expose additional anon THP stats to userspace for improved tuning.

   - "mm: finish isolate/putback_lru_page()" from Kefeng Wang: more
     folio conversions and removal of now-unused page-based APIs.

   - "replace per-quota region priorities histogram buffer with
     per-context one" from SeongJae Park. DAMON histogram
     rationalization.

   - "Docs/damon: update GitHub repo URLs and maintainer-profile" from
     SeongJae Park. DAMON documentation updates.

   - "mm/vdpa: correct misuse of non-direct-reclaim __GFP_NOFAIL and
     improve related doc and warn" from Jason Wang: fixes usage of page
     allocator __GFP_NOFAIL and GFP_ATOMIC flags.

   - "mm: split underused THPs" from Yu Zhao. Improve THP=always policy.
     This was overprovisioning THPs in sparsely accessed memory areas.

   - "zram: introduce custom comp backends API" frm Sergey Senozhatsky.
     Add support for zram run-time compression algorithm tuning.

   - "mm: Care about shadow stack guard gap when getting an unmapped
     area" from Mark Brown. Fix up the various arch_get_unmapped_area()
     implementations to better respect guard areas.

   - "Improve mem_cgroup_iter()" from Kinsey Ho. Improve the reliability
     of mem_cgroup_iter() and various code cleanups.

   - "mm: Support huge pfnmaps" from Peter Xu. Extends the usage of huge
     pfnmap support.

   - "resource: Fix region_intersects() vs add_memory_driver_managed()"
     from Huang Ying. Fix a bug in region_intersects() for systems with
     CXL memory.

   - "mm: hwpoison: two more poison recovery" from Kefeng Wang. Teaches
     a couple more code paths to correctly recover from the encountering
     of poisoned memry.

   - "mm: enable large folios swap-in support" from Barry Song. Support
     the swapin of mTHP memory into appropriately-sized folios, rather
     than into single-page folios"

* tag 'mm-stable-2024-09-20-02-31' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm: (416 commits)
  zram: free secondary algorithms names
  uprobes: turn xol_area->pages[2] into xol_area->page
  uprobes: introduce the global struct vm_special_mapping xol_mapping
  Revert "uprobes: use vm_special_mapping close() functionality"
  mm: support large folios swap-in for sync io devices
  mm: add nr argument in mem_cgroup_swapin_uncharge_swap() helper to support large folios
  mm: fix swap_read_folio_zeromap() for large folios with partial zeromap
  mm/debug_vm_pgtable: Use pxdp_get() for accessing page table entries
  set_memory: add __must_check to generic stubs
  mm/vma: return the exact errno in vms_gather_munmap_vmas()
  memcg: cleanup with !CONFIG_MEMCG_V1
  mm/show_mem.c: report alloc tags in human readable units
  mm: support poison recovery from copy_present_page()
  mm: support poison recovery from do_cow_fault()
  resource, kunit: add test case for region_intersects()
  resource: make alloc_free_mem_region() works for iomem_resource
  mm: z3fold: deprecate CONFIG_Z3FOLD
  vfio/pci: implement huge_fault support
  mm/arm64: support large pfn mappings
  mm/x86: support large pfn mappings
  ...
2024-09-21 07:29:05 -07:00

1990 lines
45 KiB
C

// SPDX-License-Identifier: GPL-2.0
#define _GNU_SOURCE
#include <linux/mman.h>
#include <sys/mman.h>
#include <stdint.h>
#include <asm-generic/unistd.h>
#include <string.h>
#include <sys/time.h>
#include <sys/resource.h>
#include <stdbool.h>
#include "../kselftest.h"
#include <syscall.h>
#include <errno.h>
#include <stdio.h>
#include <stdlib.h>
#include <fcntl.h>
#include <sys/ioctl.h>
#include <sys/vfs.h>
#include <sys/stat.h>
#include "mseal_helpers.h"
static unsigned long get_vma_size(void *addr, int *prot)
{
FILE *maps;
char line[256];
int size = 0;
uintptr_t addr_start, addr_end;
char protstr[5];
*prot = 0;
maps = fopen("/proc/self/maps", "r");
if (!maps)
return 0;
while (fgets(line, sizeof(line), maps)) {
if (sscanf(line, "%lx-%lx %4s", &addr_start, &addr_end, protstr) == 3) {
if (addr_start == (uintptr_t) addr) {
size = addr_end - addr_start;
if (protstr[0] == 'r')
*prot |= 0x4;
if (protstr[1] == 'w')
*prot |= 0x2;
if (protstr[2] == 'x')
*prot |= 0x1;
break;
}
}
}
fclose(maps);
return size;
}
/*
* define sys_xyx to call syscall directly.
*/
static int sys_mseal(void *start, size_t len)
{
int sret;
errno = 0;
sret = syscall(__NR_mseal, start, len, 0);
return sret;
}
static int sys_mprotect(void *ptr, size_t size, unsigned long prot)
{
int sret;
errno = 0;
sret = syscall(__NR_mprotect, ptr, size, prot);
return sret;
}
static int sys_mprotect_pkey(void *ptr, size_t size, unsigned long orig_prot,
unsigned long pkey)
{
int sret;
errno = 0;
sret = syscall(__NR_pkey_mprotect, ptr, size, orig_prot, pkey);
return sret;
}
static int sys_munmap(void *ptr, size_t size)
{
int sret;
errno = 0;
sret = syscall(__NR_munmap, ptr, size);
return sret;
}
static int sys_madvise(void *start, size_t len, int types)
{
int sret;
errno = 0;
sret = syscall(__NR_madvise, start, len, types);
return sret;
}
static void *sys_mremap(void *addr, size_t old_len, size_t new_len,
unsigned long flags, void *new_addr)
{
void *sret;
errno = 0;
sret = (void *) syscall(__NR_mremap, addr, old_len, new_len, flags, new_addr);
return sret;
}
static int sys_pkey_alloc(unsigned long flags, unsigned long init_val)
{
int ret = syscall(__NR_pkey_alloc, flags, init_val);
return ret;
}
static unsigned int __read_pkey_reg(void)
{
unsigned int pkey_reg = 0;
#if defined(__i386__) || defined(__x86_64__) /* arch */
unsigned int eax, edx;
unsigned int ecx = 0;
asm volatile(".byte 0x0f,0x01,0xee\n\t"
: "=a" (eax), "=d" (edx)
: "c" (ecx));
pkey_reg = eax;
#endif
return pkey_reg;
}
static void __write_pkey_reg(u64 pkey_reg)
{
#if defined(__i386__) || defined(__x86_64__) /* arch */
unsigned int eax = pkey_reg;
unsigned int ecx = 0;
unsigned int edx = 0;
asm volatile(".byte 0x0f,0x01,0xef\n\t"
: : "a" (eax), "c" (ecx), "d" (edx));
#endif
}
static unsigned long pkey_bit_position(int pkey)
{
return pkey * PKEY_BITS_PER_PKEY;
}
static u64 set_pkey_bits(u64 reg, int pkey, u64 flags)
{
unsigned long shift = pkey_bit_position(pkey);
/* mask out bits from pkey in old value */
reg &= ~((u64)PKEY_MASK << shift);
/* OR in new bits for pkey */
reg |= (flags & PKEY_MASK) << shift;
return reg;
}
static void set_pkey(int pkey, unsigned long pkey_value)
{
u64 new_pkey_reg;
new_pkey_reg = set_pkey_bits(__read_pkey_reg(), pkey, pkey_value);
__write_pkey_reg(new_pkey_reg);
}
static void setup_single_address(int size, void **ptrOut)
{
void *ptr;
ptr = mmap(NULL, size, PROT_READ, MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
*ptrOut = ptr;
}
static void setup_single_address_rw(int size, void **ptrOut)
{
void *ptr;
unsigned long mapflags = MAP_ANONYMOUS | MAP_PRIVATE;
ptr = mmap(NULL, size, PROT_READ | PROT_WRITE, mapflags, -1, 0);
*ptrOut = ptr;
}
static int clean_single_address(void *ptr, int size)
{
int ret;
ret = munmap(ptr, size);
return ret;
}
static int seal_single_address(void *ptr, int size)
{
int ret;
ret = sys_mseal(ptr, size);
return ret;
}
bool seal_support(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
ptr = mmap(NULL, page_size, PROT_READ, MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
if (ptr == (void *) -1)
return false;
ret = sys_mseal(ptr, page_size);
if (ret < 0)
return false;
return true;
}
bool pkey_supported(void)
{
#if defined(__i386__) || defined(__x86_64__) /* arch */
int pkey = sys_pkey_alloc(0, 0);
if (pkey > 0)
return true;
#endif
return false;
}
static void test_seal_addseal(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_unmapped_start(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* munmap 2 pages from ptr. */
ret = sys_munmap(ptr, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
/* mprotect will fail because 2 pages from ptr are unmapped. */
ret = sys_mprotect(ptr, size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(ret < 0);
/* mseal will fail because 2 pages from ptr are unmapped. */
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(ret < 0);
ret = sys_mseal(ptr + 2 * page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_unmapped_middle(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* munmap 2 pages from ptr + page. */
ret = sys_munmap(ptr + page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
/* mprotect will fail, since middle 2 pages are unmapped. */
ret = sys_mprotect(ptr, size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(ret < 0);
/* mseal will fail as well. */
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(ret < 0);
/* we still can add seal to the first page and last page*/
ret = sys_mseal(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_mseal(ptr + 3 * page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_unmapped_end(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* unmap last 2 pages. */
ret = sys_munmap(ptr + 2 * page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
/* mprotect will fail since last 2 pages are unmapped. */
ret = sys_mprotect(ptr, size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(ret < 0);
/* mseal will fail as well. */
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(ret < 0);
/* The first 2 pages is not sealed, and can add seals */
ret = sys_mseal(ptr, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_multiple_vmas(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split the vma into 3. */
ret = sys_mprotect(ptr + page_size, 2 * page_size,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* mprotect will get applied to all 4 pages - 3 VMAs. */
ret = sys_mprotect(ptr, size, PROT_READ);
FAIL_TEST_IF_FALSE(!ret);
/* use mprotect to split the vma into 3. */
ret = sys_mprotect(ptr + page_size, 2 * page_size,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* mseal get applied to all 4 pages - 3 VMAs. */
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_split_start(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split at middle */
ret = sys_mprotect(ptr, 2 * page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* seal the first page, this will split the VMA */
ret = sys_mseal(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
/* add seal to the remain 3 pages */
ret = sys_mseal(ptr + page_size, 3 * page_size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_split_end(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split at middle */
ret = sys_mprotect(ptr, 2 * page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* seal the last page */
ret = sys_mseal(ptr + 3 * page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
/* Adding seals to the first 3 pages */
ret = sys_mseal(ptr, 3 * page_size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_invalid_input(void)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(8 * page_size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
ret = clean_single_address(ptr + 4 * page_size, 4 * page_size);
FAIL_TEST_IF_FALSE(!ret);
/* invalid flag */
ret = syscall(__NR_mseal, ptr, size, 0x20);
FAIL_TEST_IF_FALSE(ret < 0);
/* unaligned address */
ret = sys_mseal(ptr + 1, 2 * page_size);
FAIL_TEST_IF_FALSE(ret < 0);
/* length too big */
ret = sys_mseal(ptr, 5 * page_size);
FAIL_TEST_IF_FALSE(ret < 0);
/* length overflow */
ret = sys_mseal(ptr, UINT64_MAX/page_size);
FAIL_TEST_IF_FALSE(ret < 0);
/* start is not in a valid VMA */
ret = sys_mseal(ptr - page_size, 5 * page_size);
FAIL_TEST_IF_FALSE(ret < 0);
REPORT_TEST_PASS();
}
static void test_seal_zero_length(void)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
ret = sys_mprotect(ptr, 0, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* seal 0 length will be OK, same as mprotect */
ret = sys_mseal(ptr, 0);
FAIL_TEST_IF_FALSE(!ret);
/* verify the 4 pages are not sealed by previous call. */
ret = sys_mprotect(ptr, size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_zero_address(void)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
int prot;
/* use mmap to change protection. */
ptr = mmap(0, size, PROT_NONE,
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
FAIL_TEST_IF_FALSE(ptr == 0);
size = get_vma_size(ptr, &prot);
FAIL_TEST_IF_FALSE(size == 4 * page_size);
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
/* verify the 4 pages are sealed by previous call. */
ret = sys_mprotect(ptr, size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(ret);
REPORT_TEST_PASS();
}
static void test_seal_twice(void)
{
int ret;
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
/* apply the same seal will be OK. idempotent. */
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_mprotect(ptr, size, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_start_mprotect(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* the first page is sealed. */
ret = sys_mprotect(ptr, page_size, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
/* pages after the first page is not sealed. */
ret = sys_mprotect(ptr + page_size, page_size * 3,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_end_mprotect(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr + page_size, 3 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* first page is not sealed */
ret = sys_mprotect(ptr, page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* last 3 page are sealed */
ret = sys_mprotect(ptr + page_size, page_size * 3,
PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_unalign_len(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr, page_size * 2 - 1);
FAIL_TEST_IF_FALSE(!ret);
}
/* 2 pages are sealed. */
ret = sys_mprotect(ptr, page_size * 2, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_mprotect(ptr + page_size * 2, page_size,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_unalign_len_variant_2(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr, page_size * 2 + 1);
FAIL_TEST_IF_FALSE(!ret);
}
/* 3 pages are sealed. */
ret = sys_mprotect(ptr, page_size * 3, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_mprotect(ptr + page_size * 3, page_size,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_two_vma(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split */
ret = sys_mprotect(ptr, page_size * 2, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
ret = seal_single_address(ptr, page_size * 4);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_mprotect(ptr, page_size * 2, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_mprotect(ptr + page_size * 2, page_size * 2,
PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_two_vma_with_split(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split as two vma. */
ret = sys_mprotect(ptr, page_size * 2, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* mseal can apply across 2 vma, also split them. */
if (seal) {
ret = seal_single_address(ptr + page_size, page_size * 2);
FAIL_TEST_IF_FALSE(!ret);
}
/* the first page is not sealed. */
ret = sys_mprotect(ptr, page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* the second page is sealed. */
ret = sys_mprotect(ptr + page_size, page_size, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
/* the third page is sealed. */
ret = sys_mprotect(ptr + 2 * page_size, page_size,
PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
/* the fouth page is not sealed. */
ret = sys_mprotect(ptr + 3 * page_size, page_size,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_partial_mprotect(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* seal one page. */
if (seal) {
ret = seal_single_address(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* mprotect first 2 page will fail, since the first page are sealed. */
ret = sys_mprotect(ptr, 2 * page_size, PROT_READ | PROT_WRITE);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_partial_mprotect_tail(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 2 * page_size;
int ret;
int prot;
/*
* Check if a partial mseal (that results in two vmas) works correctly.
* It might mprotect the first, but it'll never touch the second (msealed) vma.
*/
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr + page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_mprotect(ptr, size, PROT_EXEC);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
FAIL_TEST_IF_FALSE(get_vma_size(ptr + page_size, &prot) > 0);
FAIL_TEST_IF_FALSE(prot == 0x4);
}
REPORT_TEST_PASS();
}
static void test_seal_mprotect_two_vma_with_gap(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split. */
ret = sys_mprotect(ptr, page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* use mprotect to split. */
ret = sys_mprotect(ptr + 3 * page_size, page_size,
PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* use munmap to free two pages in the middle */
ret = sys_munmap(ptr + page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
/* mprotect will fail, because there is a gap in the address. */
/* notes, internally mprotect still updated the first page. */
ret = sys_mprotect(ptr, 4 * page_size, PROT_READ);
FAIL_TEST_IF_FALSE(ret < 0);
/* mseal will fail as well. */
ret = sys_mseal(ptr, 4 * page_size);
FAIL_TEST_IF_FALSE(ret < 0);
/* the first page is not sealed. */
ret = sys_mprotect(ptr, page_size, PROT_READ);
FAIL_TEST_IF_FALSE(ret == 0);
/* the last page is not sealed. */
ret = sys_mprotect(ptr + 3 * page_size, page_size, PROT_READ);
FAIL_TEST_IF_FALSE(ret == 0);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_split(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split. */
ret = sys_mprotect(ptr, page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* seal all 4 pages. */
if (seal) {
ret = sys_mseal(ptr, 4 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* mprotect is sealed. */
ret = sys_mprotect(ptr, 2 * page_size, PROT_READ);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_mprotect(ptr + 2 * page_size, 2 * page_size, PROT_READ);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_mprotect_merge(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split one page. */
ret = sys_mprotect(ptr, page_size, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
/* seal first two pages. */
if (seal) {
ret = sys_mseal(ptr, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* 2 pages are sealed. */
ret = sys_mprotect(ptr, 2 * page_size, PROT_READ);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
/* last 2 pages are not sealed. */
ret = sys_mprotect(ptr + 2 * page_size, 2 * page_size, PROT_READ);
FAIL_TEST_IF_FALSE(ret == 0);
REPORT_TEST_PASS();
}
static void test_seal_munmap(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* 4 pages are sealed. */
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
/*
* allocate 4 pages,
* use mprotect to split it as two VMAs
* seal the whole range
* munmap will fail on both
*/
static void test_seal_munmap_two_vma(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect to split */
ret = sys_mprotect(ptr, page_size * 2, PROT_READ | PROT_WRITE);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_munmap(ptr, page_size * 2);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr + page_size, page_size * 2);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
/*
* allocate a VMA with 4 pages.
* munmap the middle 2 pages.
* seal the whole 4 pages, will fail.
* munmap the first page will be OK.
* munmap the last page will be OK.
*/
static void test_seal_munmap_vma_with_gap(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
ret = sys_munmap(ptr + page_size, page_size * 2);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
/* can't have gap in the middle. */
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(ret < 0);
}
ret = sys_munmap(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr + page_size * 2, page_size);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_munmap_partial_across_vmas(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 2 * page_size;
int ret;
int prot;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr + page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
FAIL_TEST_IF_FALSE(get_vma_size(ptr + page_size, &prot) > 0);
FAIL_TEST_IF_FALSE(prot == 0x4);
}
REPORT_TEST_PASS();
}
static void test_munmap_start_freed(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
int prot;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* unmap the first page. */
ret = sys_munmap(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
/* seal the last 3 pages. */
if (seal) {
ret = sys_mseal(ptr + page_size, 3 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* unmap from the first page. */
ret = sys_munmap(ptr, size);
if (seal) {
FAIL_TEST_IF_FALSE(ret < 0);
size = get_vma_size(ptr + page_size, &prot);
FAIL_TEST_IF_FALSE(size == page_size * 3);
} else {
/* note: this will be OK, even the first page is */
/* already unmapped. */
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + page_size, &prot);
FAIL_TEST_IF_FALSE(size == 0);
}
REPORT_TEST_PASS();
}
static void test_munmap_end_freed(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* unmap last page. */
ret = sys_munmap(ptr + page_size * 3, page_size);
FAIL_TEST_IF_FALSE(!ret);
/* seal the first 3 pages. */
if (seal) {
ret = sys_mseal(ptr, 3 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* unmap all pages. */
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_munmap_middle_freed(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
int prot;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* unmap 2 pages in the middle. */
ret = sys_munmap(ptr + page_size, page_size * 2);
FAIL_TEST_IF_FALSE(!ret);
/* seal the first page. */
if (seal) {
ret = sys_mseal(ptr, page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* munmap all 4 pages. */
ret = sys_munmap(ptr, size);
if (seal) {
FAIL_TEST_IF_FALSE(ret < 0);
size = get_vma_size(ptr, &prot);
FAIL_TEST_IF_FALSE(size == page_size);
size = get_vma_size(ptr + page_size * 3, &prot);
FAIL_TEST_IF_FALSE(size == page_size);
} else {
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr, &prot);
FAIL_TEST_IF_FALSE(size == 0);
size = get_vma_size(ptr + page_size * 3, &prot);
FAIL_TEST_IF_FALSE(size == 0);
}
REPORT_TEST_PASS();
}
static void test_seal_mremap_shrink(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* shrink from 4 pages to 2 pages. */
ret2 = sys_mremap(ptr, size, 2 * page_size, 0, 0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == (void *) MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else {
FAIL_TEST_IF_FALSE(ret2 != (void *) MAP_FAILED);
}
REPORT_TEST_PASS();
}
static void test_seal_mremap_expand(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* ummap last 2 pages. */
ret = sys_munmap(ptr + 2 * page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
ret = sys_mseal(ptr, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* expand from 2 page to 4 pages. */
ret2 = sys_mremap(ptr, 2 * page_size, 4 * page_size, 0, 0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else {
FAIL_TEST_IF_FALSE(ret2 == ptr);
}
REPORT_TEST_PASS();
}
static void test_seal_mremap_move(bool seal)
{
void *ptr, *newPtr;
unsigned long page_size = getpagesize();
unsigned long size = page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
setup_single_address(size, &newPtr);
FAIL_TEST_IF_FALSE(newPtr != (void *)-1);
ret = clean_single_address(newPtr, size);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* move from ptr to fixed address. */
ret2 = sys_mremap(ptr, size, size, MREMAP_MAYMOVE | MREMAP_FIXED, newPtr);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else {
FAIL_TEST_IF_FALSE(ret2 != MAP_FAILED);
}
REPORT_TEST_PASS();
}
static void test_seal_mmap_overwrite_prot(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* use mmap to change protection. */
ret2 = mmap(ptr, size, PROT_NONE,
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else
FAIL_TEST_IF_FALSE(ret2 == ptr);
REPORT_TEST_PASS();
}
static void test_seal_mmap_expand(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 12 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* ummap last 4 pages. */
ret = sys_munmap(ptr + 8 * page_size, 4 * page_size);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
ret = sys_mseal(ptr, 8 * page_size);
FAIL_TEST_IF_FALSE(!ret);
}
/* use mmap to expand. */
ret2 = mmap(ptr, size, PROT_READ,
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else
FAIL_TEST_IF_FALSE(ret2 == ptr);
REPORT_TEST_PASS();
}
static void test_seal_mmap_shrink(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 12 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* use mmap to shrink. */
ret2 = mmap(ptr, 8 * page_size, PROT_READ,
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else
FAIL_TEST_IF_FALSE(ret2 == ptr);
REPORT_TEST_PASS();
}
static void test_seal_mremap_shrink_fixed(bool seal)
{
void *ptr;
void *newAddr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
setup_single_address(size, &newAddr);
FAIL_TEST_IF_FALSE(newAddr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* mremap to move and shrink to fixed address */
ret2 = sys_mremap(ptr, size, 2 * page_size, MREMAP_MAYMOVE | MREMAP_FIXED,
newAddr);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else
FAIL_TEST_IF_FALSE(ret2 == newAddr);
REPORT_TEST_PASS();
}
static void test_seal_mremap_expand_fixed(bool seal)
{
void *ptr;
void *newAddr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(page_size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
setup_single_address(size, &newAddr);
FAIL_TEST_IF_FALSE(newAddr != (void *)-1);
if (seal) {
ret = sys_mseal(newAddr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* mremap to move and expand to fixed address */
ret2 = sys_mremap(ptr, page_size, size, MREMAP_MAYMOVE | MREMAP_FIXED,
newAddr);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else
FAIL_TEST_IF_FALSE(ret2 == newAddr);
REPORT_TEST_PASS();
}
static void test_seal_mremap_move_fixed(bool seal)
{
void *ptr;
void *newAddr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
setup_single_address(size, &newAddr);
FAIL_TEST_IF_FALSE(newAddr != (void *)-1);
if (seal) {
ret = sys_mseal(newAddr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* mremap to move to fixed address */
ret2 = sys_mremap(ptr, size, size, MREMAP_MAYMOVE | MREMAP_FIXED, newAddr);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else
FAIL_TEST_IF_FALSE(ret2 == newAddr);
REPORT_TEST_PASS();
}
static void test_seal_mremap_move_fixed_zero(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/*
* MREMAP_FIXED can move the mapping to zero address
*/
ret2 = sys_mremap(ptr, size, 2 * page_size, MREMAP_MAYMOVE | MREMAP_FIXED,
0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else {
FAIL_TEST_IF_FALSE(ret2 == 0);
}
REPORT_TEST_PASS();
}
static void test_seal_mremap_move_dontunmap(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* mremap to move, and don't unmap src addr. */
ret2 = sys_mremap(ptr, size, size, MREMAP_MAYMOVE | MREMAP_DONTUNMAP, 0);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else {
/* kernel will allocate a new address */
FAIL_TEST_IF_FALSE(ret2 != MAP_FAILED);
}
REPORT_TEST_PASS();
}
static void test_seal_mremap_move_dontunmap_anyaddr(bool seal)
{
void *ptr, *ptr2;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
void *ret2;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/*
* The new address is any address that not allocated.
* use allocate/free to similate that.
*/
setup_single_address(size, &ptr2);
FAIL_TEST_IF_FALSE(ptr2 != (void *)-1);
ret = sys_munmap(ptr2, size);
FAIL_TEST_IF_FALSE(!ret);
/*
* remap to any address.
*/
ret2 = sys_mremap(ptr, size, size, MREMAP_MAYMOVE | MREMAP_DONTUNMAP,
(void *) ptr2);
if (seal) {
FAIL_TEST_IF_FALSE(ret2 == MAP_FAILED);
FAIL_TEST_IF_FALSE(errno == EPERM);
} else {
/* remap success and return ptr2 */
FAIL_TEST_IF_FALSE(ret2 == ptr2);
}
REPORT_TEST_PASS();
}
static void test_seal_merge_and_split(void)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size;
int ret;
int prot;
/* (24 RO) */
setup_single_address(24 * page_size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
/* use mprotect(NONE) to set out boundary */
/* (1 NONE) (22 RO) (1 NONE) */
ret = sys_mprotect(ptr, page_size, PROT_NONE);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_mprotect(ptr + 23 * page_size, page_size, PROT_NONE);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + page_size, &prot);
FAIL_TEST_IF_FALSE(size == 22 * page_size);
FAIL_TEST_IF_FALSE(prot == 4);
/* use mseal to split from beginning */
/* (1 NONE) (1 RO_SEAL) (21 RO) (1 NONE) */
ret = sys_mseal(ptr + page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + page_size, &prot);
FAIL_TEST_IF_FALSE(size == page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
size = get_vma_size(ptr + 2 * page_size, &prot);
FAIL_TEST_IF_FALSE(size == 21 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
/* use mseal to split from the end. */
/* (1 NONE) (1 RO_SEAL) (20 RO) (1 RO_SEAL) (1 NONE) */
ret = sys_mseal(ptr + 22 * page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + 22 * page_size, &prot);
FAIL_TEST_IF_FALSE(size == page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
size = get_vma_size(ptr + 2 * page_size, &prot);
FAIL_TEST_IF_FALSE(size == 20 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
/* merge with prev. */
/* (1 NONE) (2 RO_SEAL) (19 RO) (1 RO_SEAL) (1 NONE) */
ret = sys_mseal(ptr + 2 * page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + page_size, &prot);
FAIL_TEST_IF_FALSE(size == 2 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
/* merge with after. */
/* (1 NONE) (2 RO_SEAL) (18 RO) (2 RO_SEALS) (1 NONE) */
ret = sys_mseal(ptr + 21 * page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + 21 * page_size, &prot);
FAIL_TEST_IF_FALSE(size == 2 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
/* split and merge from prev */
/* (1 NONE) (3 RO_SEAL) (17 RO) (2 RO_SEALS) (1 NONE) */
ret = sys_mseal(ptr + 2 * page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + 1 * page_size, &prot);
FAIL_TEST_IF_FALSE(size == 3 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
ret = sys_munmap(ptr + page_size, page_size);
FAIL_TEST_IF_FALSE(ret < 0);
ret = sys_mprotect(ptr + 2 * page_size, page_size, PROT_NONE);
FAIL_TEST_IF_FALSE(ret < 0);
/* split and merge from next */
/* (1 NONE) (3 RO_SEAL) (16 RO) (3 RO_SEALS) (1 NONE) */
ret = sys_mseal(ptr + 20 * page_size, 2 * page_size);
FAIL_TEST_IF_FALSE(!ret);
FAIL_TEST_IF_FALSE(prot == 0x4);
size = get_vma_size(ptr + 20 * page_size, &prot);
FAIL_TEST_IF_FALSE(size == 3 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
/* merge from middle of prev and middle of next. */
/* (1 NONE) (22 RO_SEAL) (1 NONE) */
ret = sys_mseal(ptr + 2 * page_size, 20 * page_size);
FAIL_TEST_IF_FALSE(!ret);
size = get_vma_size(ptr + page_size, &prot);
FAIL_TEST_IF_FALSE(size == 22 * page_size);
FAIL_TEST_IF_FALSE(prot == 0x4);
REPORT_TEST_PASS();
}
static void test_seal_discard_ro_anon_on_rw(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address_rw(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* sealing doesn't take effect on RW memory. */
ret = sys_madvise(ptr, size, MADV_DONTNEED);
FAIL_TEST_IF_FALSE(!ret);
/* base seal still apply. */
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_discard_ro_anon_on_pkey(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
int pkey;
SKIP_TEST_IF_FALSE(pkey_supported());
setup_single_address_rw(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
pkey = sys_pkey_alloc(0, 0);
FAIL_TEST_IF_FALSE(pkey > 0);
ret = sys_mprotect_pkey((void *)ptr, size, PROT_READ | PROT_WRITE, pkey);
FAIL_TEST_IF_FALSE(!ret);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* sealing doesn't take effect if PKRU allow write. */
set_pkey(pkey, 0);
ret = sys_madvise(ptr, size, MADV_DONTNEED);
FAIL_TEST_IF_FALSE(!ret);
/* sealing will take effect if PKRU deny write. */
set_pkey(pkey, PKEY_DISABLE_WRITE);
ret = sys_madvise(ptr, size, MADV_DONTNEED);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
/* base seal still apply. */
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_discard_ro_anon_on_filebacked(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
int fd;
unsigned long mapflags = MAP_PRIVATE;
fd = memfd_create("test", 0);
FAIL_TEST_IF_FALSE(fd > 0);
ret = fallocate(fd, 0, 0, size);
FAIL_TEST_IF_FALSE(!ret);
ptr = mmap(NULL, size, PROT_READ, mapflags, fd, 0);
FAIL_TEST_IF_FALSE(ptr != MAP_FAILED);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* sealing doesn't apply for file backed mapping. */
ret = sys_madvise(ptr, size, MADV_DONTNEED);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
close(fd);
REPORT_TEST_PASS();
}
static void test_seal_discard_ro_anon_on_shared(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
unsigned long mapflags = MAP_ANONYMOUS | MAP_SHARED;
ptr = mmap(NULL, size, PROT_READ, mapflags, -1, 0);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = sys_mseal(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/* sealing doesn't apply for shared mapping. */
ret = sys_madvise(ptr, size, MADV_DONTNEED);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_discard_ro_anon(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_madvise(ptr, size, MADV_DONTNEED);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_discard_across_vmas(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 2 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr + page_size, page_size);
FAIL_TEST_IF_FALSE(!ret);
}
ret = sys_madvise(ptr, size, MADV_DONTNEED);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
static void test_seal_madvise_nodiscard(bool seal)
{
void *ptr;
unsigned long page_size = getpagesize();
unsigned long size = 4 * page_size;
int ret;
setup_single_address(size, &ptr);
FAIL_TEST_IF_FALSE(ptr != (void *)-1);
if (seal) {
ret = seal_single_address(ptr, size);
FAIL_TEST_IF_FALSE(!ret);
}
/*
* Test a random madvise flag like MADV_RANDOM that does not touch page
* contents (and thus should work for msealed VMAs). RANDOM also happens to
* share bits with other discard-ish flags like REMOVE.
*/
ret = sys_madvise(ptr, size, MADV_RANDOM);
FAIL_TEST_IF_FALSE(!ret);
ret = sys_munmap(ptr, size);
if (seal)
FAIL_TEST_IF_FALSE(ret < 0);
else
FAIL_TEST_IF_FALSE(!ret);
REPORT_TEST_PASS();
}
int main(int argc, char **argv)
{
bool test_seal = seal_support();
ksft_print_header();
if (!test_seal)
ksft_exit_skip("sealing not supported, check CONFIG_64BIT\n");
if (!pkey_supported())
ksft_print_msg("PKEY not supported\n");
ksft_set_plan(88);
test_seal_addseal();
test_seal_unmapped_start();
test_seal_unmapped_middle();
test_seal_unmapped_end();
test_seal_multiple_vmas();
test_seal_split_start();
test_seal_split_end();
test_seal_invalid_input();
test_seal_zero_length();
test_seal_twice();
test_seal_mprotect(false);
test_seal_mprotect(true);
test_seal_start_mprotect(false);
test_seal_start_mprotect(true);
test_seal_end_mprotect(false);
test_seal_end_mprotect(true);
test_seal_mprotect_unalign_len(false);
test_seal_mprotect_unalign_len(true);
test_seal_mprotect_unalign_len_variant_2(false);
test_seal_mprotect_unalign_len_variant_2(true);
test_seal_mprotect_two_vma(false);
test_seal_mprotect_two_vma(true);
test_seal_mprotect_two_vma_with_split(false);
test_seal_mprotect_two_vma_with_split(true);
test_seal_mprotect_partial_mprotect(false);
test_seal_mprotect_partial_mprotect(true);
test_seal_mprotect_two_vma_with_gap(false);
test_seal_mprotect_two_vma_with_gap(true);
test_seal_mprotect_merge(false);
test_seal_mprotect_merge(true);
test_seal_mprotect_split(false);
test_seal_mprotect_split(true);
test_seal_mprotect_partial_mprotect_tail(false);
test_seal_mprotect_partial_mprotect_tail(true);
test_seal_munmap(false);
test_seal_munmap(true);
test_seal_munmap_two_vma(false);
test_seal_munmap_two_vma(true);
test_seal_munmap_vma_with_gap(false);
test_seal_munmap_vma_with_gap(true);
test_seal_munmap_partial_across_vmas(false);
test_seal_munmap_partial_across_vmas(true);
test_munmap_start_freed(false);
test_munmap_start_freed(true);
test_munmap_middle_freed(false);
test_munmap_middle_freed(true);
test_munmap_end_freed(false);
test_munmap_end_freed(true);
test_seal_mremap_shrink(false);
test_seal_mremap_shrink(true);
test_seal_mremap_expand(false);
test_seal_mremap_expand(true);
test_seal_mremap_move(false);
test_seal_mremap_move(true);
test_seal_mremap_shrink_fixed(false);
test_seal_mremap_shrink_fixed(true);
test_seal_mremap_expand_fixed(false);
test_seal_mremap_expand_fixed(true);
test_seal_mremap_move_fixed(false);
test_seal_mremap_move_fixed(true);
test_seal_mremap_move_dontunmap(false);
test_seal_mremap_move_dontunmap(true);
test_seal_mremap_move_fixed_zero(false);
test_seal_mremap_move_fixed_zero(true);
test_seal_mremap_move_dontunmap_anyaddr(false);
test_seal_mremap_move_dontunmap_anyaddr(true);
test_seal_madvise_nodiscard(false);
test_seal_madvise_nodiscard(true);
test_seal_discard_ro_anon(false);
test_seal_discard_ro_anon(true);
test_seal_discard_across_vmas(false);
test_seal_discard_across_vmas(true);
test_seal_discard_ro_anon_on_rw(false);
test_seal_discard_ro_anon_on_rw(true);
test_seal_discard_ro_anon_on_shared(false);
test_seal_discard_ro_anon_on_shared(true);
test_seal_discard_ro_anon_on_filebacked(false);
test_seal_discard_ro_anon_on_filebacked(true);
test_seal_mmap_overwrite_prot(false);
test_seal_mmap_overwrite_prot(true);
test_seal_mmap_expand(false);
test_seal_mmap_expand(true);
test_seal_mmap_shrink(false);
test_seal_mmap_shrink(true);
test_seal_merge_and_split();
test_seal_zero_address();
test_seal_discard_ro_anon_on_pkey(false);
test_seal_discard_ro_anon_on_pkey(true);
ksft_finished();
}