1/* $NetBSD: vmpagemd.h,v 1.17 2020/12/20 16:38:26 skrll Exp $ */ 2 3/*- 4 * Copyright (c) 2011 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Raytheon BBN Technologies Corp and Defense Advanced Research Projects 9 * Agency and which was developed by Matt Thomas of 3am Software Foundry. 10 * 11 * This material is based upon work supported by the Defense Advanced Research 12 * Projects Agency and Space and Naval Warfare Systems Center, Pacific, under 13 * Contract No. N66001-09-C-2073. 14 * Approved for Public Release, Distribution Unlimited 15 * 16 * Redistribution and use in source and binary forms, with or without 17 * modification, are permitted provided that the following conditions 18 * are met: 19 * 1. Redistributions of source code must retain the above copyright 20 * notice, this list of conditions and the following disclaimer. 21 * 2. Redistributions in binary form must reproduce the above copyright 22 * notice, this list of conditions and the following disclaimer in the 23 * documentation and/or other materials provided with the distribution. 24 * 25 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 26 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 27 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 28 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 29 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 30 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 31 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 32 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 33 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 34 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 35 * POSSIBILITY OF SUCH DAMAGE. 36 */ 37 38#ifndef _UVM_PMAP_VMPAGEMD_H_ 39#define _UVM_PMAP_VMPAGEMD_H_ 40 41#ifdef _LOCORE 42#error use assym.h instead 43#endif 44 45//#ifdef _MODULE 46//#error this file should not be included by loadable kernel modules 47//#endif 48 49#ifdef _KERNEL_OPT 50#include "opt_modular.h" 51#include "opt_multiprocessor.h" 52#endif 53 54#include <sys/atomic.h> 55#include <sys/mutex.h> 56 57#define __HAVE_VM_PAGE_MD 58 59typedef struct pv_entry { 60 struct pv_entry *pv_next; 61 struct pmap *pv_pmap; 62 vaddr_t pv_va; 63#define PV_KENTER __BIT(0) 64} *pv_entry_t; 65 66#define PV_ISKENTER_P(pv) (((pv->pv_va) & PV_KENTER) != 0) 67 68#ifndef _MODULE 69 70#define VM_PAGEMD_VMPAGE __BIT(0) /* page is vm managed */ 71#define VM_PAGEMD_REFERENCED __BIT(1) /* page has been referenced */ 72#define VM_PAGEMD_MODIFIED __BIT(2) /* page has been modified */ 73#define VM_PAGEMD_POOLPAGE __BIT(3) /* page is used as a poolpage */ 74#define VM_PAGEMD_EXECPAGE __BIT(4) /* page is exec mapped */ 75#ifdef PMAP_VIRTUAL_CACHE_ALIASES 76#define VM_PAGEMD_UNCACHED __BIT(5) /* page is mapped uncached */ 77#endif 78 79#define VM_PAGEMD_VMPAGE_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_VMPAGE) != 0) 80#define VM_PAGEMD_REFERENCED_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_REFERENCED) != 0) 81#define VM_PAGEMD_MODIFIED_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_MODIFIED) != 0) 82#define VM_PAGEMD_POOLPAGE_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_POOLPAGE) != 0) 83#define VM_PAGEMD_EXECPAGE_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_EXECPAGE) != 0) 84#ifdef PMAP_VIRTUAL_CACHE_ALIASES 85#define VM_PAGEMD_CACHED_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_UNCACHED) == 0) 86#define VM_PAGEMD_UNCACHED_P(mdpg) (((mdpg)->mdpg_attrs & VM_PAGEMD_UNCACHED) != 0) 87#endif 88 89#endif /* !_MODULE */ 90 91struct vm_page_md { 92 volatile unsigned long mdpg_attrs; /* page attributes */ 93 struct pv_entry mdpg_first; /* pv_entry first */ 94#if defined(MULTIPROCESSOR) || defined(MODULAR) || defined(_MODULE) 95 kmutex_t *mdpg_lock; /* pv list lock */ 96#endif 97}; 98 99#ifndef _MODULE 100#if defined(MULTIPROCESSOR) || defined(MODULAR) 101#define VM_PAGEMD_PVLIST_LOCK_INIT(mdpg) (mdpg)->mdpg_lock = NULL 102#else 103#define VM_PAGEMD_PVLIST_LOCK_INIT(mdpg) __nothing 104#endif /* MULTIPROCESSOR || MODULAR */ 105 106#define VM_PAGEMD_PVLIST_LOCK(mdpg) pmap_pvlist_lock(mdpg, 1) 107#define VM_PAGEMD_PVLIST_READLOCK(mdpg) pmap_pvlist_lock(mdpg, 0) 108#define VM_PAGEMD_PVLIST_UNLOCK(mdpg) pmap_pvlist_unlock(mdpg) 109#define VM_PAGEMD_PVLIST_LOCKED_P(mdpg) pmap_pvlist_locked_p(mdpg) 110#define VM_PAGEMD_PVLIST_GEN(mdpg) ((mdpg)->mdpg_attrs >> 16) 111 112#define VM_PAGEMD_PVLIST_EMPTY_P(mdpg) ((mdpg)->mdpg_first.pv_pmap == NULL) 113 114#ifdef _KERNEL 115#if defined(MULTIPROCESSOR) || defined(MODULAR) 116kmutex_t *pmap_pvlist_lock_addr(struct vm_page_md *); 117#else 118extern kmutex_t pmap_pvlist_mutex; 119static __inline kmutex_t * 120pmap_pvlist_lock_addr(struct vm_page_md *mdpg) 121{ 122 return &pmap_pvlist_mutex; 123} 124#endif 125 126static __inline uintptr_t 127pmap_pvlist_lock(struct vm_page_md *mdpg, uintptr_t increment) 128{ 129 mutex_spin_enter(pmap_pvlist_lock_addr(mdpg)); 130 const uintptr_t gen = VM_PAGEMD_PVLIST_GEN(mdpg); 131 mdpg->mdpg_attrs += increment << 16; 132 return gen; 133} 134 135static __inline uintptr_t 136pmap_pvlist_unlock(struct vm_page_md *mdpg) 137{ 138 const uintptr_t gen = VM_PAGEMD_PVLIST_GEN(mdpg); 139 mutex_spin_exit(pmap_pvlist_lock_addr(mdpg)); 140 return gen; 141} 142 143static __inline bool 144pmap_pvlist_locked_p(struct vm_page_md *mdpg) 145{ 146 147 return mutex_owned(pmap_pvlist_lock_addr(mdpg)); 148} 149#endif /* _KERNEL */ 150 151#define VM_MDPAGE_INIT(pg) \ 152do { \ 153 (pg)->mdpage.mdpg_first.pv_next = NULL; \ 154 (pg)->mdpage.mdpg_first.pv_pmap = NULL; \ 155 (pg)->mdpage.mdpg_first.pv_va = VM_PAGE_TO_PHYS(pg); \ 156 (pg)->mdpage.mdpg_attrs = VM_PAGEMD_VMPAGE; \ 157 VM_PAGEMD_PVLIST_LOCK_INIT(&(pg)->mdpage); \ 158} while (/* CONSTCOND */ 0) 159 160#endif /* _MODULE */ 161 162#endif /* _UVM_PMAP_VMPAGEMD_H_ */ 163