Searched refs:page (Results 201 - 225 of 3148) sorted by last modified time

1234567891011>>

/linux-master/arch/x86/kernel/cpu/sgx/
H A Dmain.c51 * from the input list, and made available for the page allocator. SECS pages
60 struct sgx_epc_page *page; local
69 page = list_first_entry(dirty_page_list, struct sgx_epc_page, list);
72 * Checking page->poison without holding the node->lock
75 * called for a page that sgx_free_epc_page() will put onto
78 if (page->poison) {
79 struct sgx_epc_section *section = &sgx_epc_sections[page->section];
83 list_move(&page->list, &node->sgx_poison_page_list);
89 ret = __eremove(sgx_get_epc_virt_addr(page));
92 * page i
112 struct sgx_encl_page *page = epc_page->owner; local
144 struct sgx_encl_page *page = epc_page->owner; local
445 struct sgx_epc_page *page = NULL; local
476 struct sgx_epc_page *page; local
507 sgx_mark_page_reclaimable(struct sgx_epc_page *page) argument
525 sgx_unmark_page_reclaimable(struct sgx_epc_page *page) argument
562 struct sgx_epc_page *page; local
603 sgx_free_epc_page(struct sgx_epc_page *page) argument
679 struct sgx_epc_page *page = sgx_paddr_to_page(pfn << PAGE_SHIFT); local
[all...]
/linux-master/arch/x86/kernel/
H A Damd_gart_64.c225 static dma_addr_t gart_map_page(struct device *dev, struct page *page, argument
231 phys_addr_t paddr = page_to_phys(page) + offset;
375 * Merge chunks that have page aligned sizes into a continuous mapping.
409 * page boundary and the new one doesn't have an
767 * Unmap the IOMMU part of the GART. The alias of the page is
797 * Set unmapped entries to a scratch page instead of 0.
803 panic("Cannot allocate iommu scratch page");
/linux-master/arch/x86/include/asm/
H A Dpgtable.h6 #include <asm/page.h>
10 * Macro to mark a page protection value as UC-
51 * ZERO_PAGE is a global shared page that is always zero: used
61 extern struct mm_struct *pgd_page_get_mm(struct page *page);
271 /* NOTE: when predicate huge page, consider also pmd_devmap, or use pmd_leaf */
902 * the kernel copy of the page tables.
1025 * Conversion functions: convert a page and protection to a page entry,
1026 * and a page entr
[all...]
H A Dpgtable_64.h12 * the x86-64 page table tree.
173 * Conversion functions: convert a page and protection to a page entry,
174 * and a page entry and page directory to the page they refer to.
202 * remember soft dirty over page migration
259 #define vmemmap ((struct page *)VMEMMAP_START)
H A Dio.h41 #include <asm/page.h>
155 * Change "struct page" to physical address.
157 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-master/arch/x86/entry/vdso/
H A DMakefile36 -z max-page-size=4096
89 -z max-page-size=4096
/linux-master/arch/sparc/vdso/
H A DMakefile26 -z max-page-size=8192
/linux-master/arch/sparc/mm/
H A Dtlb.c120 struct page *page; local
126 page = pfn_to_page(pfn);
127 if (PageReserved(page))
130 /* A real file page? */
131 folio = page_folio(page);
136 paddr = (unsigned long) page_address(page);
180 * for huge zero page. Huge zero pages are like hugetlb
182 * for TSB entries. So, huge zero page counts go into
/linux-master/arch/sparc/include/asm/
H A Dpgtable_64.h3 * pgtable.h: SpitFire page table operations.
13 * the SpitFire page tables.
23 #include <asm/page.h>
27 * The page copy blockops can use 0x6000000 to 0x8000000.
48 /* PMD_SHIFT determines the size of the area a second-level page
56 /* PUD_SHIFT determines the size of the area a third-level page
64 /* PGDIR_SHIFT determines what a fourth-level page table entry can map */
71 #error MAX_PHYS_ADDRESS_BITS exceeds what kernel page tables can support
86 #define vmemmap ((struct page *)VMEMMAP_BASE)
93 /* Entries per page director
[all...]
/linux-master/arch/sh/mm/
H A Dcache.c60 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, argument
64 struct folio *folio = page_folio(page);
68 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK);
78 flush_cache_page(vma, vaddr, page_to_pfn(page));
81 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, argument
85 struct folio *folio = page_folio(page);
89 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK);
99 void copy_user_highpage(struct page *t
128 clear_user_highpage(struct page *page, unsigned long vaddr) argument
158 __flush_anon_page(struct page *page, unsigned long vmaddr) argument
243 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument
[all...]
H A Dmmap.c15 #include <asm/page.h>
43 * To avoid cache aliases, we map the shared page with same color.
H A Dcache-sh4.c107 * Write back & invalidate the D-cache of the page.
209 * Write back and invalidate I/D-caches for the page.
212 * PFN: Physical page number
218 struct page *page; local
229 page = pfn_to_page(pfn);
237 /* If the page isn't present, there is nothing to do here. */
244 struct folio *folio = page_folio(page);
251 page_mapped(page));
253 vaddr = kmap_coherent(page, addres
[all...]
/linux-master/arch/sh/include/asm/
H A Dhugetlb.h6 #include <asm/page.h>
/linux-master/arch/s390/mm/
H A Dhugetlbpage.c124 struct page *page; local
132 page = pud_page(__pud(rste));
136 page = pmd_page(__pmd(rste));
141 if (!test_and_set_bit(PG_arch_1, &page->flags))
H A Dfault.c272 * This routine handles page faults. It determines the address,
495 struct page *page; local
537 page = follow_page(vma, addr, FOLL_WRITE | FOLL_GET);
538 if (IS_ERR_OR_NULL(page)) {
542 if (arch_make_page_accessible(page))
544 put_page(page);
548 page = phys_to_page(addr);
549 if (unlikely(!try_get_page(page)))
551 rc = arch_make_page_accessible(page);
[all...]
/linux-master/arch/s390/kernel/
H A Dipl.c210 char *page) \
212 return scnprintf(page, PAGE_SIZE, _format, ##args); \
314 char *page)
316 return sprintf(page, "%s\n", ipl_type_str(ipl_info.type));
322 struct kobj_attribute *attr, char *page)
324 return sprintf(page, "%i\n", !!ipl_secure_flag);
331 struct kobj_attribute *attr, char *page)
333 return sprintf(page, "%i\n", !!sclp.has_sipl);
340 struct kobj_attribute *attr, char *page)
346 return sprintf(page, "
313 ipl_type_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
321 ipl_secure_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
330 ipl_has_secure_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
339 ipl_vm_parm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
352 sys_ipl_device_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
544 ipl_ccw_loadparm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
702 reipl_generic_vmparm_show(struct ipl_parameter_block *ipb, char *page) argument
744 reipl_nss_vmparm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
758 reipl_ccw_vmparm_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
844 reipl_generic_loadparm_show(struct ipl_parameter_block *ipb, char *page) argument
903 reipl_fcp_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
1005 reipl_nvme_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
1026 reipl_ccw_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
1132 reipl_eckd_clear_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
1159 reipl_nss_name_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
1247 reipl_type_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
1735 dump_type_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
2075 on_reboot_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
2101 on_panic_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
2127 on_restart_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
2163 on_halt_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
2189 on_poff_show(struct kobject *kobj, struct kobj_attribute *attr, char *page) argument
[all...]
/linux-master/arch/s390/include/asm/
H A Dpgtable.h17 #include <linux/page-flags.h>
23 #include <asm/page.h>
47 * The S390 doesn't have any external MMU info: the kernel page
55 * ZERO_PAGE is a global shared page that is always zero; used
91 extern struct page *__bootdata_preserved(vmemmap);
123 * P Page-Protection Bit: Store access not possible for page
133 * P Page-Protection Bit: Store access not possible for page
167 /* Hardware bits in the page table entry */
173 /* Software bits in the page table entry */
179 #define _PAGE_SPECIAL 0x040 /* SW associated with special page */
1381 mk_pte(struct page *page, pgprot_t pgprot) argument
[all...]
H A Dhugetlb.h13 #include <asm/page.h>
82 static inline pte_t mk_huge_pte(struct page *page, pgprot_t pgprot) argument
84 return mk_pte(page, pgprot);
H A Dio.h14 #include <asm/page.h>
/linux-master/arch/riscv/mm/
H A Ddma-noncoherent.c117 void arch_dma_prep_coherent(struct page *page, size_t size) argument
119 void *flush_addr = page_address(page);
123 noncoherent_cache_ops.wback_inv(page_to_phys(page), size);
/linux-master/arch/riscv/include/asm/
H A Dpgtable.h28 /* Number of entries in the page global directory */
30 /* Number of entries in the page table */
34 * Half of the kernel address space (1/4 of the entries of the page global
90 #define vmemmap ((struct page *)VMEMMAP_START - (phys_ram_base >> PAGE_SHIFT))
119 #include <asm/page.h>
270 static inline struct page *pmd_page(pmd_t pmd)
323 /* Yields the page frame number (PFN) of a page table entry */
336 /* Constructs a page table entry */
346 #define mk_pte(page, pro
[all...]
H A Dhugetlb.h6 #include <asm/page.h>
/linux-master/arch/powerpc/platforms/pseries/
H A Dvio.c34 #include <asm/page.h>
515 static dma_addr_t vio_dma_iommu_map_page(struct device *dev, struct page *page, argument
526 ret = iommu_map_page(dev, tbl, page, offset, size, dma_get_mask(dev),
/linux-master/arch/powerpc/platforms/ps3/
H A Dsystem-bus.c504 * to the dma address (mapping) of the first page.
555 * comprises a page address and offset into that page. The dma_addr_t
556 * returned will point to the same byte within the page as was passed in.
559 static dma_addr_t ps3_sb_map_page(struct device *_dev, struct page *page, argument
566 void *ptr = page_address(page) + offset;
581 static dma_addr_t ps3_ioc0_map_page(struct device *_dev, struct page *page, argument
590 void *ptr = page_address(page)
[all...]
/linux-master/arch/powerpc/mm/
H A Dpgtable_64.c34 #include <asm/page.h>
56 * page table size
91 struct page *vmemmap;
100 /* 4 level page table */
101 struct page *p4d_page(p4d_t p4d)
112 struct page *pud_page(pud_t pud)
124 * For PTE page, we have a PTE_FRAG_SIZE (4K) aligned virtual address.
126 struct page *pmd_page(pmd_t pmd)

Completed in 473 milliseconds

1234567891011>>