/linux-master/arch/arm64/mm/ |
H A D | mteswap.c | 23 int mte_save_tags(struct page *page) argument 27 if (!page_mte_tagged(page)) 34 mte_save_page_tags(page_address(page), tag_storage); 36 /* lookup the swap entry.val from the page */ 37 ret = xa_store(&mte_pages, page_swap_entry(page).val, tag_storage, 50 void mte_restore_tags(swp_entry_t entry, struct page *page) argument 57 if (try_page_mte_tagging(page)) { 58 mte_restore_page_tags(page_address(page), tag [all...] |
/linux-master/mm/ |
H A D | debug_page_alloc.c | 3 #include <linux/page-isolation.h> 35 bool __set_page_guard(struct zone *zone, struct page *page, unsigned int order, argument 41 __SetPageGuard(page); 42 INIT_LIST_HEAD(&page->buddy_list); 43 set_page_private(page, order); 51 void __clear_page_guard(struct zone *zone, struct page *page, unsigned int order, argument 54 __ClearPageGuard(page); 56 set_page_private(page, [all...] |
H A D | migrate_device.c | 74 struct page *page; local 82 page = pmd_page(*pmdp); 83 if (is_huge_zero_page(page)) { 89 get_page(page); 91 if (unlikely(!trylock_page(page))) 94 ret = split_huge_page(page); 95 unlock_page(page); 96 put_page(page); 111 struct page *pag local 325 migrate_vma_check_page(struct page *page, struct page *fault_page) argument 371 struct page *page = migrate_pfn_to_page(src_pfns[i]); local 418 struct page *page = migrate_pfn_to_page(src_pfns[i]); local 565 migrate_vma_insert_page(struct migrate_vma *migrate, unsigned long addr, struct page *page, unsigned long *src) argument 695 struct page *page = migrate_pfn_to_page(src_pfns[i]); local 823 struct page *page = migrate_pfn_to_page(src_pfns[i]); local 904 struct page *page = pfn_to_page(pfn); local 931 migrate_device_coherent_page(struct page *page) argument [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | cacheflush_32.h | 5 #include <linux/page-flags.h> 20 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 22 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 25 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 27 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 38 void sparc_flush_page_to_ram(struct page *page); 43 static inline void flush_dcache_page(struct page *page) argument 45 flush_dcache_folio(page_folio(page)); [all...] |
H A D | cacheflush_64.h | 5 #include <asm/page.h> 26 #define flush_cache_page(vma, page, pfn) \ 51 static inline void flush_dcache_page(struct page *page) argument 53 flush_dcache_folio(page_folio(page)); 56 void flush_ptrace_access(struct vm_area_struct *, struct page *, 60 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 62 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 64 flush_ptrace_access(vma, page, vaddr, src, len, 0); \ 67 #define copy_from_user_page(vma, page, vadd [all...] |
/linux-master/arch/nios2/include/asm/ |
H A D | page.h | 5 * MMU support based on asm/page.h from mips which is: 22 * PAGE_SHIFT determines the page size 29 * PAGE_OFFSET -- the first address of the first page of memory. 48 #define clear_page(page) memset((page), 0, PAGE_SIZE) 51 struct page; 53 extern void clear_user_page(void *addr, unsigned long vaddr, struct page *page); 55 struct page *to); 60 typedef struct page *pgtable_ [all...] |
/linux-master/tools/testing/selftests/bpf/ |
H A D | bpf_arena_alloc.h | 24 void __arena *page = page_frag_cur_page[cpu]; local 31 if (!page) { 33 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); 34 if (!page) 36 cast_kern(page); 37 page_frag_cur_page[cpu] = page; 39 obj_cnt = page + PAGE_SIZE - 8; 42 cast_kern(page); 43 obj_cnt = page + PAGE_SIZE - 8; 52 return page [all...] |
/linux-master/mm/kmsan/ |
H A D | shadow.c | 22 #define shadow_page_for(page) ((page)->kmsan_shadow) 24 #define origin_page_for(page) ((page)->kmsan_origin) 26 static void *shadow_ptr_for(struct page *page) argument 28 return page_address(shadow_page_for(page)); 31 static void *origin_ptr_for(struct page *page) argument 33 return page_address(origin_page_for(page)); 36 page_has_metadata(struct page *page) argument 41 set_no_shadow_origin_page(struct page *page) argument 127 struct page *page; local 172 kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags) argument 207 kmsan_free_page(struct page *page, unsigned int order) argument 281 struct page *page; local 304 kmsan_setup_meta(struct page *page, struct page *shadow, struct page *origin, int order) argument [all...] |
/linux-master/fs/jfs/ |
H A D | jfs_metapage.c | 26 uint pagealloc; /* # of page allocations */ 27 uint pagefree; /* # of page frees */ 49 unlock_page(mp->page); 51 lock_page(mp->page); 59 * Must have mp->page locked 80 #define mp_anchor(page) ((struct meta_anchor *)page_private(page)) 82 static inline struct metapage *page_to_mp(struct page *page, int offset) argument 84 if (!PagePrivate(page)) 89 insert_metapage(struct page *page, struct metapage *mp) argument 116 remove_metapage(struct page *page, struct metapage *mp) argument 135 inc_io(struct page *page) argument 140 dec_io(struct page *page, void (*handler) (struct page *)) argument 147 page_to_mp(struct page *page, int offset) argument 152 insert_metapage(struct page *page, struct metapage *mp) argument 162 remove_metapage(struct page *page, struct metapage *mp) argument 221 drop_metapage(struct page *page, struct metapage *mp) argument 260 last_read_complete(struct page *page) argument 269 struct page *page = bio->bi_private; local 302 last_write_complete(struct page *page) argument 324 struct page *page = bio->bi_private; local 336 metapage_writepage(struct page *page, struct writeback_control *wbc) argument 472 struct page *page = &folio->page; local 582 struct page *page; local 694 metapage_write_one(struct page *page) argument 725 struct page *page = mp->page; local 759 struct page *page = mp->page; local 801 struct page *page; local [all...] |
/linux-master/include/linux/ |
H A D | rmap.h | 19 * an anonymous page pointing to this anon_vma needs to be unmapped: 23 * in mprotect), the mapping field of an anonymous page cannot point 28 * the anon_vma object itself: we're guaranteed no page can be 36 * guarantee that the vma of page tables will exist for 180 * No special request: A mapped anonymous (sub)page is possibly shared between 185 /* The anonymous (sub)page is exclusive to a single process. */ 198 struct page *page, int nr_pages, enum rmap_level level) 213 VM_WARN_ON_FOLIO(page_folio(page) != folio, folio); 214 VM_WARN_ON_FOLIO(page_folio(page 197 __folio_rmap_sanity_checks(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 318 __folio_dup_file_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 345 folio_dup_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages) argument 362 folio_dup_file_rmap_pmd(struct folio *folio, struct page *page) argument 372 __folio_try_dup_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma, enum rmap_level level) argument 445 folio_try_dup_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma) argument 476 folio_try_dup_anon_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *src_vma) argument 488 __folio_try_share_anon_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 581 folio_try_share_anon_rmap_pte(struct folio *folio, struct page *page) argument 610 folio_try_share_anon_rmap_pmd(struct folio *folio, struct page *page) argument 758 page_mkclean(struct page *page) argument [all...] |
H A D | set_memory.h | 28 static inline int set_direct_map_invalid_noflush(struct page *page) argument 32 static inline int set_direct_map_default_noflush(struct page *page) argument 37 static inline bool kernel_page_present(struct page *page) argument
|
/linux-master/fs/ufs/ |
H A D | dir.c | 16 * Migration to usage of "page cache" on May 2006 by 45 static void ufs_commit_chunk(struct page *page, loff_t pos, unsigned len) argument 47 struct address_space *mapping = page->mapping; 51 block_write_end(NULL, mapping, pos, len, len, page, NULL); 56 unlock_page(page); 69 static inline void ufs_put_page(struct page *page) argument 71 kunmap(page); 72 put_page(page); 79 struct page *page; local 91 ufs_set_link(struct inode *dir, struct ufs_dir_entry *de, struct page *page, struct inode *inode, bool update_times) argument 116 ufs_check_page(struct page *page) argument 198 struct page *page = read_mapping_page(mapping, n, NULL); local 237 struct page *page = ufs_get_page(dir, 0); local 265 struct page *page = NULL; local 320 struct page *page = NULL; local 451 struct page *page = ufs_get_page(inode, n); local 502 ufs_delete_entry(struct inode *inode, struct ufs_dir_entry *dir, struct page * page) argument 555 struct page *page = grab_cache_page(mapping, 0); local 603 struct page *page = NULL; local [all...] |
/linux-master/fs/qnx6/ |
H A D | dir.c | 27 static struct page *qnx6_get_page(struct inode *dir, unsigned long n) 30 struct page *page = read_mapping_page(mapping, n, NULL); local 31 if (!IS_ERR(page)) 32 kmap(page); 33 return page; 47 struct page **p) 52 /* within page */ 55 struct page *page local 70 struct page *page; local 128 struct page *page = qnx6_get_page(inode, n); local 180 struct page *page; local 218 struct page *page = NULL; local [all...] |
H A D | namei.c | 20 struct page *page; local 28 ino = qnx6_find_entry(len, dir, name, &page); 31 qnx6_put_page(page);
|
/linux-master/net/core/ |
H A D | page_pool.c | 18 #include <linux/page-flags.h> 66 * page_pool_get_stats() - fetch page pool stats 67 * @pool: pool from which page was allocated 198 * DMA_BIDIRECTIONAL is for allowing page used for DMA sending, 208 /* In order to request DMA-sync-for-device the page 230 /* For system page pool instance we use a singular stats object 232 * (also percpu) page pool instance. 271 * page_pool_create_percpu() - create a page pool for a given cpu. 305 * page_pool_create() - create a page pool 314 static void page_pool_return_page(struct page_pool *pool, struct page *pag 320 struct page *page; local 372 struct page *page; local 386 page_pool_dma_sync_for_device(struct page_pool *pool, struct page *page, unsigned int dma_sync_size) argument 398 page_pool_dma_map(struct page_pool *pool, struct page *page) argument 430 page_pool_set_pp_info(struct page_pool *pool, struct page *page) argument 447 page_pool_clear_pp_info(struct page *page) argument 456 struct page *page; local 486 struct page *page; local 541 struct page *page; local 579 __page_pool_release_page_dma(struct page_pool *pool, struct page *page) argument 603 page_pool_return_page(struct page_pool *pool, struct page *page) argument 624 page_pool_recycle_in_ring(struct page_pool *pool, struct page *page) argument 646 page_pool_recycle_in_cache(struct page *page, struct page_pool *pool) argument 660 __page_pool_page_can_be_recycled(const struct page *page) argument 672 __page_pool_put_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) argument 719 page_pool_put_unrefed_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) argument 753 struct page *page = virt_to_head_page(data[i]); local 792 page_pool_drain_frag(struct page_pool *pool, struct page *page) argument 815 struct page *page = pool->frag_page; local 830 struct page *page = pool->frag_page; local 872 struct page *page; local 897 struct page *page; local 1014 struct page *page; local [all...] |
/linux-master/drivers/nvdimm/ |
H A D | pmem.h | 4 #include <linux/page-flags.h> 37 static inline bool test_and_clear_pmem_poison(struct page *page) argument 39 return TestClearPageHWPoison(page); 42 static inline bool test_and_clear_pmem_poison(struct page *page) argument
|
/linux-master/arch/riscv/include/asm/ |
H A D | set_memory.h | 43 int set_direct_map_invalid_noflush(struct page *page); 44 int set_direct_map_default_noflush(struct page *page); 45 bool kernel_page_present(struct page *page);
|
/linux-master/arch/m68k/include/asm/ |
H A D | page_no.h | 10 #define clear_page(page) memset((page), 0, PAGE_SIZE) 13 #define clear_user_page(page, vaddr, pg) clear_page(page) 33 #define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET))
|
H A D | virtconvert.h | 14 #include <asm/page.h> 31 /* Permanent address of a page. */ 32 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
/linux-master/arch/parisc/mm/ |
H A D | ioremap.c | 32 struct page *page; local 37 for (page = virt_to_page(t_addr); 38 page <= virt_to_page(t_end); page++) { 39 if(!PageReserved(page))
|
/linux-master/arch/microblaze/mm/ |
H A D | consistent.c | 18 void arch_dma_prep_coherent(struct page *page, size_t size) argument 20 phys_addr_t paddr = page_to_phys(page);
|
/linux-master/fs/squashfs/ |
H A D | symlink.c | 35 struct page *page = &folio->page; local 36 struct inode *inode = page->mapping->host; 39 int index = page->index << PAGE_SHIFT; 47 TRACE("Entered squashfs_symlink_readpage, page index %ld, start block " 48 "%llx, offset %x\n", page->index, block, offset); 67 * kmap_atomic to map the page. Instead call the underlying 81 pageaddr = kmap_atomic(page); 92 flush_dcache_page(page); [all...] |
H A D | file_cache.c | 20 /* Read separately compressed datablock and memcopy into page cache */ 21 int squashfs_readpage_block(struct page *page, u64 block, int bsize, int expected) argument 23 struct inode *i = page->mapping->host; 29 ERROR("Unable to read page, block %llx, size %x\n", block, 32 squashfs_copy_cache(page, buffer, expected, 0);
|
/linux-master/arch/s390/include/asm/ |
H A D | kfence.h | 9 #include <asm/page.h> 11 void __kernel_map_pages(struct page *page, int numpages, int enable); 22 * but earlier where page table allocations still happen with memblock.
|
/linux-master/arch/arm/include/asm/ |
H A D | page-nommu.h | 3 * arch/arm/include/asm/page-nommu.h 11 #define clear_page(page) memset((page), 0, PAGE_SIZE) 14 #define clear_user_page(page, vaddr, pg) clear_page(page)
|