freebsd-src/sys/arm64/include/vmparam.h
Mark Johnston 4441dd4094 vm_phys: Fix a typo
Fixes:	b16b4c22d2 ("vm_page: Implement lazy page initialization")
Reported by:	Steffen Nurpmeso <steffen@sdaoden.eu>
2024-06-16 13:33:00 -04:00

327 lines
10 KiB
C

/*-
* Copyright (c) 1990 The Regents of the University of California.
* All rights reserved.
* Copyright (c) 1994 John S. Dyson
* All rights reserved.
*
* This code is derived from software contributed to Berkeley by
* William Jolitz.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. Neither the name of the University nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
* from: FreeBSD: src/sys/i386/include/vmparam.h,v 1.33 2000/03/30
*/
#ifdef __arm__
#include <arm/vmparam.h>
#else /* !__arm__ */
#ifndef _MACHINE_VMPARAM_H_
#define _MACHINE_VMPARAM_H_
/*
* Virtual memory related constants, all in bytes
*/
#ifndef MAXTSIZ
#define MAXTSIZ (1*1024*1024*1024) /* max text size */
#endif
#ifndef DFLDSIZ
#define DFLDSIZ (128*1024*1024) /* initial data size limit */
#endif
#ifndef MAXDSIZ
#define MAXDSIZ (1*1024*1024*1024) /* max data size */
#endif
#ifndef DFLSSIZ
#define DFLSSIZ (128*1024*1024) /* initial stack size limit */
#endif
#ifndef MAXSSIZ
#define MAXSSIZ (1*1024*1024*1024) /* max stack size */
#endif
#ifndef SGROWSIZ
#define SGROWSIZ (128*1024) /* amount to grow stack */
#endif
/*
* The physical address space is sparsely populated.
*/
#define VM_PHYSSEG_SPARSE
/*
* The number of PHYSSEG entries.
*/
#define VM_PHYSSEG_MAX 64
/*
* Create three free page pools: VM_FREEPOOL_DEFAULT is the default pool from
* which physical pages are allocated and VM_FREEPOOL_DIRECT is the pool from
* which physical pages for page tables and small UMA objects are allocated.
* VM_FREEPOOL_LAZYINIT is a special-purpose pool that is populated only during
* boot and is used to implement deferred initialization of page structures.
*/
#define VM_NFREEPOOL 3
#define VM_FREEPOOL_LAZYINIT 0
#define VM_FREEPOOL_DEFAULT 1
#define VM_FREEPOOL_DIRECT 2
/*
* Create two free page lists: VM_FREELIST_DMA32 is for physical pages that have
* physical addresses below 4G, and VM_FREELIST_DEFAULT is for all other
* physical pages.
*/
#define VM_NFREELIST 2
#define VM_FREELIST_DEFAULT 0
#define VM_FREELIST_DMA32 1
/*
* When PAGE_SIZE is 4KB, an allocation size of 16MB is supported in order
* to optimize the use of the direct map by UMA. Specifically, a 64-byte
* cache line contains at most 8 L2 BLOCK entries, collectively mapping 16MB
* of physical memory. By reducing the number of distinct 16MB "pages" that
* are used by UMA, the physical memory allocator reduces the likelihood of
* both 2MB page TLB misses and cache misses during the page table walk when
* a 2MB page TLB miss does occur.
*
* When PAGE_SIZE is 16KB, an allocation size of 32MB is supported. This
* size is used by level 0 reservations and L2 BLOCK mappings.
*/
#if PAGE_SIZE == PAGE_SIZE_4K
#define VM_NFREEORDER 13
#elif PAGE_SIZE == PAGE_SIZE_16K
#define VM_NFREEORDER 12
#else
#error Unsupported page size
#endif
/*
* Enable superpage reservations: 1 level.
*/
#ifndef VM_NRESERVLEVEL
#define VM_NRESERVLEVEL 1
#endif
/*
* Level 0 reservations consist of 512 pages when PAGE_SIZE is 4KB, and
* 2048 pages when PAGE_SIZE is 16KB.
*/
#ifndef VM_LEVEL_0_ORDER
#if PAGE_SIZE == PAGE_SIZE_4K
#define VM_LEVEL_0_ORDER 9
#elif PAGE_SIZE == PAGE_SIZE_16K
#define VM_LEVEL_0_ORDER 11
#else
#error Unsupported page size
#endif
#endif
/**
* Address space layout.
*
* ARMv8 implements up to a 48 bit virtual address space. The address space is
* split into 2 regions at each end of the 64 bit address space, with an
* out of range "hole" in the middle.
*
* We use the full 48 bits for each region, however the kernel may only use
* a limited range within this space.
*
* Upper region: 0xffffffffffffffff Top of virtual memory
*
* 0xfffffeffffffffff End of DMAP
* 0xffffa00000000000 Start of DMAP
*
* 0xffff027fffffffff End of KMSAN origin map
* 0xffff020000000000 Start of KMSAN origin map
*
* 0xffff017fffffffff End of KMSAN shadow map
* 0xffff010000000000 Start of KMSAN shadow map
*
* 0xffff009fffffffff End of KASAN shadow map
* 0xffff008000000000 Start of KASAN shadow map
*
* 0xffff007fffffffff End of KVA
* 0xffff000000000000 Kernel base address & start of KVA
*
* Hole: 0xfffeffffffffffff
* 0x0001000000000000
*
* Lower region: 0x0000ffffffffffff End of user address space
* 0x0000000000000000 Start of user address space
*
* We use the upper region for the kernel, and the lower region for userland.
*
* We define some interesting address constants:
*
* VM_MIN_ADDRESS and VM_MAX_ADDRESS define the start and end of the entire
* 64 bit address space, mostly just for convenience.
*
* VM_MIN_KERNEL_ADDRESS and VM_MAX_KERNEL_ADDRESS define the start and end of
* mappable kernel virtual address space.
*
* VM_MIN_USER_ADDRESS and VM_MAX_USER_ADDRESS define the start and end of the
* user address space.
*/
#define VM_MIN_ADDRESS (0x0000000000000000UL)
#define VM_MAX_ADDRESS (0xffffffffffffffffUL)
/* 512 GiB of kernel addresses */
#define VM_MIN_KERNEL_ADDRESS (0xffff000000000000UL)
#define VM_MAX_KERNEL_ADDRESS (0xffff008000000000UL)
/* 128 GiB KASAN shadow map */
#define KASAN_MIN_ADDRESS (0xffff008000000000UL)
#define KASAN_MAX_ADDRESS (0xffff00a000000000UL)
/* 512GiB KMSAN shadow map */
#define KMSAN_SHAD_MIN_ADDRESS (0xffff010000000000UL)
#define KMSAN_SHAD_MAX_ADDRESS (0xffff018000000000UL)
/* 512GiB KMSAN origin map */
#define KMSAN_ORIG_MIN_ADDRESS (0xffff020000000000UL)
#define KMSAN_ORIG_MAX_ADDRESS (0xffff028000000000UL)
/* The address bits that hold a pointer authentication code */
#define PAC_ADDR_MASK (0xff7f000000000000UL)
/* If true addr is in the kernel address space */
#define ADDR_IS_KERNEL(addr) (((addr) & (1ul << 55)) == (1ul << 55))
/* If true addr is in its canonical form (i.e. no TBI, PAC, etc.) */
#define ADDR_IS_CANONICAL(addr) \
(((addr) & 0xffff000000000000UL) == 0 || \
((addr) & 0xffff000000000000UL) == 0xffff000000000000UL)
#define ADDR_MAKE_CANONICAL(addr) ({ \
__typeof(addr) _tmp_addr = (addr); \
\
_tmp_addr &= ~0xffff000000000000UL; \
if (ADDR_IS_KERNEL(addr)) \
_tmp_addr |= 0xffff000000000000UL; \
\
_tmp_addr; \
})
/* 95 TiB maximum for the direct map region */
#define DMAP_MIN_ADDRESS (0xffffa00000000000UL)
#define DMAP_MAX_ADDRESS (0xffffff0000000000UL)
#define DMAP_MIN_PHYSADDR (dmap_phys_base)
#define DMAP_MAX_PHYSADDR (dmap_phys_max)
/*
* Checks to see if a physical address is in the DMAP range.
* - PHYS_IN_DMAP_RANGE will return true that may be within the DMAP range
* but not accessible through the DMAP, e.g. device memory between two
* DMAP physical address regions.
* - PHYS_IN_DMAP will check if DMAP address is mapped before returning true.
*
* PHYS_IN_DMAP_RANGE should only be used when a check on the address is
* performed, e.g. by checking the physical address is within phys_avail,
* or checking the virtual address is mapped.
*/
#define PHYS_IN_DMAP_RANGE(pa) ((pa) >= DMAP_MIN_PHYSADDR && \
(pa) < DMAP_MAX_PHYSADDR)
#define PHYS_IN_DMAP(pa) (PHYS_IN_DMAP_RANGE(pa) && \
pmap_klookup(PHYS_TO_DMAP(pa), NULL))
/* True if va is in the dmap range */
#define VIRT_IN_DMAP(va) ((va) >= DMAP_MIN_ADDRESS && \
(va) < (dmap_max_addr))
#define PMAP_HAS_DMAP 1
#define PHYS_TO_DMAP(pa) \
({ \
KASSERT(PHYS_IN_DMAP_RANGE(pa), \
("%s: PA out of range, PA: 0x%lx", __func__, \
(vm_paddr_t)(pa))); \
((pa) - dmap_phys_base) + DMAP_MIN_ADDRESS; \
})
#define DMAP_TO_PHYS(va) \
({ \
KASSERT(VIRT_IN_DMAP(va), \
("%s: VA out of range, VA: 0x%lx", __func__, \
(vm_offset_t)(va))); \
((va) - DMAP_MIN_ADDRESS) + dmap_phys_base; \
})
#define VM_MIN_USER_ADDRESS (0x0000000000000000UL)
#define VM_MAX_USER_ADDRESS (0x0001000000000000UL)
#define VM_MINUSER_ADDRESS (VM_MIN_USER_ADDRESS)
#define VM_MAXUSER_ADDRESS (VM_MAX_USER_ADDRESS)
#define KERNBASE (VM_MIN_KERNEL_ADDRESS)
#define SHAREDPAGE (VM_MAXUSER_ADDRESS - PAGE_SIZE)
#define USRSTACK SHAREDPAGE
/*
* How many physical pages per kmem arena virtual page.
*/
#ifndef VM_KMEM_SIZE_SCALE
#define VM_KMEM_SIZE_SCALE (1)
#endif
/*
* Optional ceiling (in bytes) on the size of the kmem arena: 60% of the
* kernel map.
*/
#ifndef VM_KMEM_SIZE_MAX
#define VM_KMEM_SIZE_MAX ((VM_MAX_KERNEL_ADDRESS - \
VM_MIN_KERNEL_ADDRESS + 1) * 3 / 5)
#endif
/*
* Initial pagein size of beginning of executable file.
*/
#ifndef VM_INITIAL_PAGEIN
#define VM_INITIAL_PAGEIN 16
#endif
#if !defined(KASAN) && !defined(KMSAN)
#define UMA_USE_DMAP
#endif
#ifndef LOCORE
extern vm_paddr_t dmap_phys_base;
extern vm_paddr_t dmap_phys_max;
extern vm_offset_t dmap_max_addr;
#endif
#define ZERO_REGION_SIZE (64 * 1024) /* 64KB */
#define DEVMAP_MAX_VADDR VM_MAX_KERNEL_ADDRESS
/*
* The pmap can create non-transparent large page mappings.
*/
#define PMAP_HAS_LARGEPAGES 1
/*
* Need a page dump array for minidump.
*/
#define MINIDUMP_PAGE_TRACKING 1
#define MINIDUMP_STARTUP_PAGE_TRACKING 1
#endif /* !_MACHINE_VMPARAM_H_ */
#endif /* !__arm__ */