/linux-master/mm/ |
H A D | zsmalloc.c | 48 #include <linux/pgtable.h>
|
H A D | vmalloc.c | 40 #include <linux/pgtable.h>
|
H A D | pgtable-generic.c | 3 * mm/pgtable-generic.c 5 * Generic pgtable methods declared in linux/pgtable.h 12 #include <linux/pgtable.h> 166 pgtable_t pgtable) 172 INIT_LIST_HEAD(&pgtable->lru); 174 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); 175 pmd_huge_pte(mm, pmdp) = pgtable; 183 pgtable_t pgtable; local 188 pgtable 165 pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp, pgtable_t pgtable) argument 247 pte_free_defer(struct mm_struct *mm, pgtable_t pgtable) argument [all...] |
H A D | mprotect.c | 31 #include <linux/pgtable.h> 325 * Populate the pgtable underneath for whatever reason if requested. 326 * When {pte|pmd|...}_alloc() failed we treat it the same way as pgtable
|
H A D | mm_init.c | 26 #include <linux/pgtable.h>
|
H A D | khugepaged.c | 1080 pgtable_t pgtable; local 1209 pgtable = pmd_pgtable(_pmd); 1218 pgtable_trans_huge_deposit(mm, pmd, pgtable); 1507 /* Keep pmd pgtable for uffd-wp; see comment in retract_page_tables() */
|
H A D | hugetlb_vmemmap.c | 13 #include <linux/pgtable.h> 55 pte_t *pgtable; local 57 pgtable = pte_alloc_one_kernel(&init_mm); 58 if (!pgtable) 61 pmd_populate_kernel(&init_mm, &__pmd, pgtable); 84 pmd_populate_kernel(&init_mm, pmd, pgtable); 88 pte_free_kernel(&init_mm, pgtable);
|
H A D | huge_memory.c | 923 pgtable_t pgtable; local 939 pgtable = pte_alloc_one(vma->vm_mm); 940 if (unlikely(!pgtable)) { 967 pte_free(vma->vm_mm, pgtable); 977 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); 992 if (pgtable) 993 pte_free(vma->vm_mm, pgtable); 1035 static void set_huge_zero_folio(pgtable_t pgtable, struct mm_struct *mm, argument 1044 pgtable_trans_huge_deposit(mm, pmd, pgtable); 1067 pgtable_t pgtable; local 1114 insert_pfn_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd, pfn_t pfn, pgprot_t prot, bool write, pgtable_t pgtable) argument 1176 pgtable_t pgtable = NULL; local 1349 pgtable_t pgtable = NULL; local 1806 pgtable_t pgtable; local 1938 pgtable_t pgtable; local 2327 pgtable_t pgtable; local 2370 pgtable_t pgtable; local [all...] |
H A D | debug_vm_pgtable.c | 25 #include <linux/pgtable.h>
|
H A D | Makefile | 39 pgtable-generic.o rmap.o vmalloc.o
|
/linux-master/kernel/events/ |
H A D | core.c | 55 #include <linux/pgtable.h>
|
/linux-master/kernel/bpf/ |
H A D | syscall.c | 30 #include <linux/pgtable.h>
|
/linux-master/include/linux/ |
H A D | pgtable.h | 6 #include <asm/pgtable.h> 38 * defined in their respective <asm/pgtable.h>. 47 * via their respective <asm/pgtable.h>. 119 void pte_free_defer(struct mm_struct *mm, pgtable_t pgtable); 758 * consecutive in the pgtable. 923 pgtable_t pgtable); 1125 * prototypes must be defined in the arch-specific asm/pgtable.h file. 1517 * or fails to copy pgtable during duplicate vm area. 1846 * pXd_leaf() is the API to check whether a pgtable entry is a huge page 1851 * A leaf pgtable entr [all...] |
H A D | page-flags.h | 1043 PAGE_TYPE_OPS(Table, table, pgtable)
|
H A D | mm.h | 30 #include <linux/pgtable.h> 118 * This macro should be defined within <asm/pgtable.h>. 130 * define their own version of this macro in <asm/pgtable.h>
|
H A D | hugetlb.h | 14 #include <linux/pgtable.h> 208 * huge_pte_offset(): Walk the hugetlb pgtable until the last level PTE. 216 * Since this function will walk all the pgtable pages (including not only 217 * high-level pgtable page, but also PUD entry that can be unshared 227 * pgtable page can go away from under us! It can be done by a pmd 241 * a concurrent pmd unshare, but it makes sure the pgtable page is safe to
|
H A D | dma-map-ops.h | 10 #include <linux/pgtable.h>
|
/linux-master/include/asm-generic/ |
H A D | io.h | 1048 #include <linux/pgtable.h>
|
/linux-master/fs/ |
H A D | dax.c | 1209 pgtable_t pgtable = NULL; local 1225 pgtable = pte_alloc_one(vma->vm_mm); 1226 if (!pgtable) 1236 if (pgtable) { 1237 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); 1248 if (pgtable) 1249 pte_free(vma->vm_mm, pgtable);
|
/linux-master/drivers/iommu/ |
H A D | io-pgtable-dart.c | 7 * Based on io-pgtable-arm. 14 #define pr_fmt(fmt) "dart io-pgtable: " fmt 19 #include <linux/io-pgtable.h>
|
H A D | io-pgtable-arm.c | 10 #define pr_fmt(fmt) "arm-lpae io-pgtable: " fmt 14 #include <linux/io-pgtable.h> 23 #include "io-pgtable-arm.h" 1193 pr_err("selftest: failed to allocate io pgtable ops\n");
|
H A D | exynos-iommu.c | 180 static sysmmu_pte_t *section_entry(sysmmu_pte_t *pgtable, sysmmu_iova_t iova) argument 182 return pgtable + lv1ent_offset(iova); 250 sysmmu_pte_t *pgtable; /* lv1 page table, 16KB */ member in struct:exynos_iommu_domain 253 spinlock_t pgtablelock; /* lock for modifying page table @ pgtable */ 300 phys_addr_t pgtable; /* assigned page table structure */ member in struct:sysmmu_drvdata 557 dev_dbg(data->sysmmu, "Page table base: %pa\n", &data->pgtable); 558 ent = section_entry(phys_to_virt(data->pgtable), fault->addr); 656 __sysmmu_set_ptbase(data, data->pgtable); 905 domain->pgtable = iommu_alloc_pages(GFP_KERNEL, 2); 906 if (!domain->pgtable) [all...] |
/linux-master/drivers/iommu/intel/ |
H A D | iommu.c | 750 struct dma_pte *pgtable; local 780 pgtable = phys_to_virt(ctx_entry->lo & VTD_PAGE_MASK); 811 pgtable = phys_to_virt(pte->val[2] & VTD_PAGE_MASK); 814 pgtable = phys_to_virt(pte->val[0] & VTD_PAGE_MASK); 818 pgtable_walk(iommu, addr >> VTD_PAGE_SHIFT, bus, devfn, pgtable, level); 1683 * in-flight DMA and copied pgtable, but there is no unmapping
|
/linux-master/drivers/iommu/arm/arm-smmu/ |
H A D | arm-smmu.h | 19 #include <linux/io-pgtable.h>
|
H A D | arm-smmu-qcom.c | 125 struct io_pgtable *pgtable = local 127 return &pgtable->cfg; 140 struct io_pgtable *pgtable = io_pgtable_ops_to_pgtable(smmu_domain->pgtbl_ops); local 155 cb->tcr[0] = arm_smmu_lpae_tcr(&pgtable->cfg);
|