/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | main.c | 51 * from the input list, and made available for the page allocator. SECS pages 60 struct sgx_epc_page *page; local 69 page = list_first_entry(dirty_page_list, struct sgx_epc_page, list); 72 * Checking page->poison without holding the node->lock 75 * called for a page that sgx_free_epc_page() will put onto 78 if (page->poison) { 79 struct sgx_epc_section *section = &sgx_epc_sections[page->section]; 83 list_move(&page->list, &node->sgx_poison_page_list); 89 ret = __eremove(sgx_get_epc_virt_addr(page)); 92 * page i 112 struct sgx_encl_page *page = epc_page->owner; local 144 struct sgx_encl_page *page = epc_page->owner; local 445 struct sgx_epc_page *page = NULL; local 476 struct sgx_epc_page *page; local 507 sgx_mark_page_reclaimable(struct sgx_epc_page *page) argument 525 sgx_unmark_page_reclaimable(struct sgx_epc_page *page) argument 562 struct sgx_epc_page *page; local 603 sgx_free_epc_page(struct sgx_epc_page *page) argument 679 struct sgx_epc_page *page = sgx_paddr_to_page(pfn << PAGE_SHIFT); local [all...] |
/linux-master/arch/x86/kernel/ |
H A D | amd_gart_64.c | 225 static dma_addr_t gart_map_page(struct device *dev, struct page *page, argument 231 phys_addr_t paddr = page_to_phys(page) + offset; 375 * Merge chunks that have page aligned sizes into a continuous mapping. 409 * page boundary and the new one doesn't have an 767 * Unmap the IOMMU part of the GART. The alias of the page is 797 * Set unmapped entries to a scratch page instead of 0. 803 panic("Cannot allocate iommu scratch page");
|
/linux-master/arch/x86/include/asm/ |
H A D | pgtable.h | 6 #include <asm/page.h> 10 * Macro to mark a page protection value as UC- 51 * ZERO_PAGE is a global shared page that is always zero: used 61 extern struct mm_struct *pgd_page_get_mm(struct page *page); 271 /* NOTE: when predicate huge page, consider also pmd_devmap, or use pmd_leaf */ 902 * the kernel copy of the page tables. 1025 * Conversion functions: convert a page and protection to a page entry, 1026 * and a page entr [all...] |
H A D | pgtable_64.h | 12 * the x86-64 page table tree. 173 * Conversion functions: convert a page and protection to a page entry, 174 * and a page entry and page directory to the page they refer to. 202 * remember soft dirty over page migration 259 #define vmemmap ((struct page *)VMEMMAP_START)
|
H A D | io.h | 41 #include <asm/page.h> 155 * Change "struct page" to physical address. 157 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
|
/linux-master/arch/x86/entry/vdso/ |
H A D | Makefile | 36 -z max-page-size=4096 89 -z max-page-size=4096
|
/linux-master/arch/sparc/vdso/ |
H A D | Makefile | 26 -z max-page-size=8192
|
/linux-master/arch/sparc/mm/ |
H A D | tlb.c | 120 struct page *page; local 126 page = pfn_to_page(pfn); 127 if (PageReserved(page)) 130 /* A real file page? */ 131 folio = page_folio(page); 136 paddr = (unsigned long) page_address(page); 180 * for huge zero page. Huge zero pages are like hugetlb 182 * for TSB entries. So, huge zero page counts go into
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_64.h | 3 * pgtable.h: SpitFire page table operations. 13 * the SpitFire page tables. 23 #include <asm/page.h> 27 * The page copy blockops can use 0x6000000 to 0x8000000. 48 /* PMD_SHIFT determines the size of the area a second-level page 56 /* PUD_SHIFT determines the size of the area a third-level page 64 /* PGDIR_SHIFT determines what a fourth-level page table entry can map */ 71 #error MAX_PHYS_ADDRESS_BITS exceeds what kernel page tables can support 86 #define vmemmap ((struct page *)VMEMMAP_BASE) 93 /* Entries per page director [all...] |
/linux-master/arch/sh/mm/ |
H A D | cache.c | 60 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, argument 64 struct folio *folio = page_folio(page); 68 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); 78 flush_cache_page(vma, vaddr, page_to_pfn(page)); 81 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, argument 85 struct folio *folio = page_folio(page); 89 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); 99 void copy_user_highpage(struct page *t 128 clear_user_highpage(struct page *page, unsigned long vaddr) argument 158 __flush_anon_page(struct page *page, unsigned long vmaddr) argument 243 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument [all...] |
H A D | mmap.c | 15 #include <asm/page.h> 43 * To avoid cache aliases, we map the shared page with same color.
|
H A D | cache-sh4.c | 107 * Write back & invalidate the D-cache of the page. 209 * Write back and invalidate I/D-caches for the page. 212 * PFN: Physical page number 218 struct page *page; local 229 page = pfn_to_page(pfn); 237 /* If the page isn't present, there is nothing to do here. */ 244 struct folio *folio = page_folio(page); 251 page_mapped(page)); 253 vaddr = kmap_coherent(page, addres [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | hugetlb.h | 6 #include <asm/page.h>
|
/linux-master/arch/s390/mm/ |
H A D | hugetlbpage.c | 124 struct page *page; local 132 page = pud_page(__pud(rste)); 136 page = pmd_page(__pmd(rste)); 141 if (!test_and_set_bit(PG_arch_1, &page->flags))
|
H A D | fault.c | 272 * This routine handles page faults. It determines the address, 495 struct page *page; local 537 page = follow_page(vma, addr, FOLL_WRITE | FOLL_GET); 538 if (IS_ERR_OR_NULL(page)) { 542 if (arch_make_page_accessible(page)) 544 put_page(page); 548 page = phys_to_page(addr); 549 if (unlikely(!try_get_page(page))) 551 rc = arch_make_page_accessible(page); [all...] |
/linux-master/arch/s390/kernel/ |
H A D | ipl.c | 210 char *page) \ 212 return scnprintf(page, PAGE_SIZE, _format, ##args); \ 314 char *page) 316 return sprintf(page, "%s\n", ipl_type_str(ipl_info.type)); 322 struct kobj_attribute *attr, char *page) 324 return sprintf(page, "%i\n", !!ipl_secure_flag); 331 struct kobj_attribute *attr, char *page) 333 return sprintf(page, "%i\n", !!sclp.has_sipl); 340 struct kobj_attribute *attr, char *page) 346 return sprintf(page, " 313 ipl_type_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 321 ipl_secure_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 330 ipl_has_secure_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 339 ipl_vm_parm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 352 sys_ipl_device_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 544 ipl_ccw_loadparm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 702 reipl_generic_vmparm_show(struct ipl_parameter_block *ipb, char *page) argument 744 reipl_nss_vmparm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 758 reipl_ccw_vmparm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 844 reipl_generic_loadparm_show(struct ipl_parameter_block *ipb, char *page) argument 903 reipl_fcp_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 1005 reipl_nvme_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 1026 reipl_ccw_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 1132 reipl_eckd_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 1159 reipl_nss_name_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 1247 reipl_type_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 1735 dump_type_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 2075 on_reboot_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 2101 on_panic_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 2127 on_restart_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 2163 on_halt_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument 2189 on_poff_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument [all...] |
/linux-master/arch/s390/include/asm/ |
H A D | pgtable.h | 17 #include <linux/page-flags.h> 23 #include <asm/page.h> 47 * The S390 doesn't have any external MMU info: the kernel page 55 * ZERO_PAGE is a global shared page that is always zero; used 91 extern struct page *__bootdata_preserved(vmemmap); 123 * P Page-Protection Bit: Store access not possible for page 133 * P Page-Protection Bit: Store access not possible for page 167 /* Hardware bits in the page table entry */ 173 /* Software bits in the page table entry */ 179 #define _PAGE_SPECIAL 0x040 /* SW associated with special page */ 1381 mk_pte(struct page *page, pgprot_t pgprot) argument [all...] |
H A D | hugetlb.h | 13 #include <asm/page.h> 82 static inline pte_t mk_huge_pte(struct page *page, pgprot_t pgprot) argument 84 return mk_pte(page, pgprot);
|
H A D | io.h | 14 #include <asm/page.h>
|
/linux-master/arch/riscv/mm/ |
H A D | dma-noncoherent.c | 117 void arch_dma_prep_coherent(struct page *page, size_t size) argument 119 void *flush_addr = page_address(page); 123 noncoherent_cache_ops.wback_inv(page_to_phys(page), size);
|
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable.h | 28 /* Number of entries in the page global directory */ 30 /* Number of entries in the page table */ 34 * Half of the kernel address space (1/4 of the entries of the page global 90 #define vmemmap ((struct page *)VMEMMAP_START - (phys_ram_base >> PAGE_SHIFT)) 119 #include <asm/page.h> 270 static inline struct page *pmd_page(pmd_t pmd) 323 /* Yields the page frame number (PFN) of a page table entry */ 336 /* Constructs a page table entry */ 346 #define mk_pte(page, pro [all...] |
H A D | hugetlb.h | 6 #include <asm/page.h>
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | vio.c | 34 #include <asm/page.h> 515 static dma_addr_t vio_dma_iommu_map_page(struct device *dev, struct page *page, argument 526 ret = iommu_map_page(dev, tbl, page, offset, size, dma_get_mask(dev),
|
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | system-bus.c | 504 * to the dma address (mapping) of the first page. 555 * comprises a page address and offset into that page. The dma_addr_t 556 * returned will point to the same byte within the page as was passed in. 559 static dma_addr_t ps3_sb_map_page(struct device *_dev, struct page *page, argument 566 void *ptr = page_address(page) + offset; 581 static dma_addr_t ps3_ioc0_map_page(struct device *_dev, struct page *page, argument 590 void *ptr = page_address(page) [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | pgtable_64.c | 34 #include <asm/page.h> 56 * page table size 91 struct page *vmemmap; 100 /* 4 level page table */ 101 struct page *p4d_page(p4d_t p4d) 112 struct page *pud_page(pud_t pud) 124 * For PTE page, we have a PTE_FRAG_SIZE (4K) aligned virtual address. 126 struct page *pmd_page(pmd_t pmd)
|