Searched refs:page (Results 76 - 100 of 3144) sorted by relevance

1234567891011>>

/linux-master/arch/arm64/mm/
H A Dmteswap.c23 int mte_save_tags(struct page *page) argument
27 if (!page_mte_tagged(page))
34 mte_save_page_tags(page_address(page), tag_storage);
36 /* lookup the swap entry.val from the page */
37 ret = xa_store(&mte_pages, page_swap_entry(page).val, tag_storage,
50 void mte_restore_tags(swp_entry_t entry, struct page *page) argument
57 if (try_page_mte_tagging(page)) {
58 mte_restore_page_tags(page_address(page), tag
[all...]
/linux-master/mm/
H A Ddebug_page_alloc.c3 #include <linux/page-isolation.h>
35 bool __set_page_guard(struct zone *zone, struct page *page, unsigned int order, argument
41 __SetPageGuard(page);
42 INIT_LIST_HEAD(&page->buddy_list);
43 set_page_private(page, order);
51 void __clear_page_guard(struct zone *zone, struct page *page, unsigned int order, argument
54 __ClearPageGuard(page);
56 set_page_private(page,
[all...]
H A Dmigrate_device.c74 struct page *page; local
82 page = pmd_page(*pmdp);
83 if (is_huge_zero_page(page)) {
89 get_page(page);
91 if (unlikely(!trylock_page(page)))
94 ret = split_huge_page(page);
95 unlock_page(page);
96 put_page(page);
111 struct page *pag local
325 migrate_vma_check_page(struct page *page, struct page *fault_page) argument
371 struct page *page = migrate_pfn_to_page(src_pfns[i]); local
418 struct page *page = migrate_pfn_to_page(src_pfns[i]); local
565 migrate_vma_insert_page(struct migrate_vma *migrate, unsigned long addr, struct page *page, unsigned long *src) argument
695 struct page *page = migrate_pfn_to_page(src_pfns[i]); local
823 struct page *page = migrate_pfn_to_page(src_pfns[i]); local
904 struct page *page = pfn_to_page(pfn); local
931 migrate_device_coherent_page(struct page *page) argument
[all...]
/linux-master/arch/sparc/include/asm/
H A Dcacheflush_32.h5 #include <linux/page-flags.h>
20 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
22 flush_cache_page(vma, vaddr, page_to_pfn(page));\
25 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
27 flush_cache_page(vma, vaddr, page_to_pfn(page));\
38 void sparc_flush_page_to_ram(struct page *page);
43 static inline void flush_dcache_page(struct page *page) argument
45 flush_dcache_folio(page_folio(page));
[all...]
H A Dcacheflush_64.h5 #include <asm/page.h>
26 #define flush_cache_page(vma, page, pfn) \
51 static inline void flush_dcache_page(struct page *page) argument
53 flush_dcache_folio(page_folio(page));
56 void flush_ptrace_access(struct vm_area_struct *, struct page *,
60 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
62 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
64 flush_ptrace_access(vma, page, vaddr, src, len, 0); \
67 #define copy_from_user_page(vma, page, vadd
[all...]
/linux-master/arch/nios2/include/asm/
H A Dpage.h5 * MMU support based on asm/page.h from mips which is:
22 * PAGE_SHIFT determines the page size
29 * PAGE_OFFSET -- the first address of the first page of memory.
48 #define clear_page(page) memset((page), 0, PAGE_SIZE)
51 struct page;
53 extern void clear_user_page(void *addr, unsigned long vaddr, struct page *page);
55 struct page *to);
60 typedef struct page *pgtable_
[all...]
/linux-master/tools/testing/selftests/bpf/
H A Dbpf_arena_alloc.h24 void __arena *page = page_frag_cur_page[cpu]; local
31 if (!page) {
33 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0);
34 if (!page)
36 cast_kern(page);
37 page_frag_cur_page[cpu] = page;
39 obj_cnt = page + PAGE_SIZE - 8;
42 cast_kern(page);
43 obj_cnt = page + PAGE_SIZE - 8;
52 return page
[all...]
/linux-master/mm/kmsan/
H A Dshadow.c22 #define shadow_page_for(page) ((page)->kmsan_shadow)
24 #define origin_page_for(page) ((page)->kmsan_origin)
26 static void *shadow_ptr_for(struct page *page) argument
28 return page_address(shadow_page_for(page));
31 static void *origin_ptr_for(struct page *page) argument
33 return page_address(origin_page_for(page));
36 page_has_metadata(struct page *page) argument
41 set_no_shadow_origin_page(struct page *page) argument
127 struct page *page; local
172 kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags) argument
207 kmsan_free_page(struct page *page, unsigned int order) argument
281 struct page *page; local
304 kmsan_setup_meta(struct page *page, struct page *shadow, struct page *origin, int order) argument
[all...]
/linux-master/fs/jfs/
H A Djfs_metapage.c26 uint pagealloc; /* # of page allocations */
27 uint pagefree; /* # of page frees */
49 unlock_page(mp->page);
51 lock_page(mp->page);
59 * Must have mp->page locked
80 #define mp_anchor(page) ((struct meta_anchor *)page_private(page))
82 static inline struct metapage *page_to_mp(struct page *page, int offset) argument
84 if (!PagePrivate(page))
89 insert_metapage(struct page *page, struct metapage *mp) argument
116 remove_metapage(struct page *page, struct metapage *mp) argument
135 inc_io(struct page *page) argument
140 dec_io(struct page *page, void (*handler) (struct page *)) argument
147 page_to_mp(struct page *page, int offset) argument
152 insert_metapage(struct page *page, struct metapage *mp) argument
162 remove_metapage(struct page *page, struct metapage *mp) argument
221 drop_metapage(struct page *page, struct metapage *mp) argument
260 last_read_complete(struct page *page) argument
269 struct page *page = bio->bi_private; local
302 last_write_complete(struct page *page) argument
324 struct page *page = bio->bi_private; local
336 metapage_writepage(struct page *page, struct writeback_control *wbc) argument
472 struct page *page = &folio->page; local
582 struct page *page; local
694 metapage_write_one(struct page *page) argument
725 struct page *page = mp->page; local
759 struct page *page = mp->page; local
801 struct page *page; local
[all...]
/linux-master/include/linux/
H A Drmap.h19 * an anonymous page pointing to this anon_vma needs to be unmapped:
23 * in mprotect), the mapping field of an anonymous page cannot point
28 * the anon_vma object itself: we're guaranteed no page can be
36 * guarantee that the vma of page tables will exist for
180 * No special request: A mapped anonymous (sub)page is possibly shared between
185 /* The anonymous (sub)page is exclusive to a single process. */
198 struct page *page, int nr_pages, enum rmap_level level)
213 VM_WARN_ON_FOLIO(page_folio(page) != folio, folio);
214 VM_WARN_ON_FOLIO(page_folio(page
197 __folio_rmap_sanity_checks(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
318 __folio_dup_file_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
345 folio_dup_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages) argument
362 folio_dup_file_rmap_pmd(struct folio *folio, struct page *page) argument
372 __folio_try_dup_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma, enum rmap_level level) argument
445 folio_try_dup_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma) argument
476 folio_try_dup_anon_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *src_vma) argument
488 __folio_try_share_anon_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
581 folio_try_share_anon_rmap_pte(struct folio *folio, struct page *page) argument
610 folio_try_share_anon_rmap_pmd(struct folio *folio, struct page *page) argument
758 page_mkclean(struct page *page) argument
[all...]
H A Dset_memory.h28 static inline int set_direct_map_invalid_noflush(struct page *page) argument
32 static inline int set_direct_map_default_noflush(struct page *page) argument
37 static inline bool kernel_page_present(struct page *page) argument
/linux-master/fs/ufs/
H A Ddir.c16 * Migration to usage of "page cache" on May 2006 by
45 static void ufs_commit_chunk(struct page *page, loff_t pos, unsigned len) argument
47 struct address_space *mapping = page->mapping;
51 block_write_end(NULL, mapping, pos, len, len, page, NULL);
56 unlock_page(page);
69 static inline void ufs_put_page(struct page *page) argument
71 kunmap(page);
72 put_page(page);
79 struct page *page; local
91 ufs_set_link(struct inode *dir, struct ufs_dir_entry *de, struct page *page, struct inode *inode, bool update_times) argument
116 ufs_check_page(struct page *page) argument
198 struct page *page = read_mapping_page(mapping, n, NULL); local
237 struct page *page = ufs_get_page(dir, 0); local
265 struct page *page = NULL; local
320 struct page *page = NULL; local
451 struct page *page = ufs_get_page(inode, n); local
502 ufs_delete_entry(struct inode *inode, struct ufs_dir_entry *dir, struct page * page) argument
555 struct page *page = grab_cache_page(mapping, 0); local
603 struct page *page = NULL; local
[all...]
/linux-master/fs/qnx6/
H A Ddir.c27 static struct page *qnx6_get_page(struct inode *dir, unsigned long n)
30 struct page *page = read_mapping_page(mapping, n, NULL); local
31 if (!IS_ERR(page))
32 kmap(page);
33 return page;
47 struct page **p)
52 /* within page */
55 struct page *page local
70 struct page *page; local
128 struct page *page = qnx6_get_page(inode, n); local
180 struct page *page; local
218 struct page *page = NULL; local
[all...]
H A Dnamei.c20 struct page *page; local
28 ino = qnx6_find_entry(len, dir, name, &page);
31 qnx6_put_page(page);
/linux-master/net/core/
H A Dpage_pool.c18 #include <linux/page-flags.h>
66 * page_pool_get_stats() - fetch page pool stats
67 * @pool: pool from which page was allocated
198 * DMA_BIDIRECTIONAL is for allowing page used for DMA sending,
208 /* In order to request DMA-sync-for-device the page
230 /* For system page pool instance we use a singular stats object
232 * (also percpu) page pool instance.
271 * page_pool_create_percpu() - create a page pool for a given cpu.
305 * page_pool_create() - create a page pool
314 static void page_pool_return_page(struct page_pool *pool, struct page *pag
320 struct page *page; local
372 struct page *page; local
386 page_pool_dma_sync_for_device(struct page_pool *pool, struct page *page, unsigned int dma_sync_size) argument
398 page_pool_dma_map(struct page_pool *pool, struct page *page) argument
430 page_pool_set_pp_info(struct page_pool *pool, struct page *page) argument
447 page_pool_clear_pp_info(struct page *page) argument
456 struct page *page; local
486 struct page *page; local
541 struct page *page; local
579 __page_pool_release_page_dma(struct page_pool *pool, struct page *page) argument
603 page_pool_return_page(struct page_pool *pool, struct page *page) argument
624 page_pool_recycle_in_ring(struct page_pool *pool, struct page *page) argument
646 page_pool_recycle_in_cache(struct page *page, struct page_pool *pool) argument
660 __page_pool_page_can_be_recycled(const struct page *page) argument
672 __page_pool_put_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) argument
719 page_pool_put_unrefed_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) argument
753 struct page *page = virt_to_head_page(data[i]); local
792 page_pool_drain_frag(struct page_pool *pool, struct page *page) argument
815 struct page *page = pool->frag_page; local
830 struct page *page = pool->frag_page; local
872 struct page *page; local
897 struct page *page; local
1014 struct page *page; local
[all...]
/linux-master/drivers/nvdimm/
H A Dpmem.h4 #include <linux/page-flags.h>
37 static inline bool test_and_clear_pmem_poison(struct page *page) argument
39 return TestClearPageHWPoison(page);
42 static inline bool test_and_clear_pmem_poison(struct page *page) argument
/linux-master/arch/riscv/include/asm/
H A Dset_memory.h43 int set_direct_map_invalid_noflush(struct page *page);
44 int set_direct_map_default_noflush(struct page *page);
45 bool kernel_page_present(struct page *page);
/linux-master/arch/m68k/include/asm/
H A Dpage_no.h10 #define clear_page(page) memset((page), 0, PAGE_SIZE)
13 #define clear_user_page(page, vaddr, pg) clear_page(page)
33 #define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET))
H A Dvirtconvert.h14 #include <asm/page.h>
31 /* Permanent address of a page. */
32 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
/linux-master/arch/parisc/mm/
H A Dioremap.c32 struct page *page; local
37 for (page = virt_to_page(t_addr);
38 page <= virt_to_page(t_end); page++) {
39 if(!PageReserved(page))
/linux-master/arch/microblaze/mm/
H A Dconsistent.c18 void arch_dma_prep_coherent(struct page *page, size_t size) argument
20 phys_addr_t paddr = page_to_phys(page);
/linux-master/fs/squashfs/
H A Dsymlink.c35 struct page *page = &folio->page; local
36 struct inode *inode = page->mapping->host;
39 int index = page->index << PAGE_SHIFT;
47 TRACE("Entered squashfs_symlink_readpage, page index %ld, start block "
48 "%llx, offset %x\n", page->index, block, offset);
67 * kmap_atomic to map the page. Instead call the underlying
81 pageaddr = kmap_atomic(page);
92 flush_dcache_page(page);
[all...]
H A Dfile_cache.c20 /* Read separately compressed datablock and memcopy into page cache */
21 int squashfs_readpage_block(struct page *page, u64 block, int bsize, int expected) argument
23 struct inode *i = page->mapping->host;
29 ERROR("Unable to read page, block %llx, size %x\n", block,
32 squashfs_copy_cache(page, buffer, expected, 0);
/linux-master/arch/s390/include/asm/
H A Dkfence.h9 #include <asm/page.h>
11 void __kernel_map_pages(struct page *page, int numpages, int enable);
22 * but earlier where page table allocations still happen with memblock.
/linux-master/arch/arm/include/asm/
H A Dpage-nommu.h3 * arch/arm/include/asm/page-nommu.h
11 #define clear_page(page) memset((page), 0, PAGE_SIZE)
14 #define clear_user_page(page, vaddr, pg) clear_page(page)

Completed in 278 milliseconds

1234567891011>>