/linux-master/tools/virtio/linux/ |
H A D | mm_types.h | 2 struct page page; member in struct:folio
|
/linux-master/arch/openrisc/include/asm/ |
H A D | cacheflush.h | 24 extern void local_dcache_page_flush(struct page *page); 25 extern void local_icache_page_inv(struct page *page); 33 #define dcache_page_flush(page) local_dcache_page_flush(page) 34 #define icache_page_inv(page) local_icache_page_inv(page) 36 #define dcache_page_flush(page) local_dcache_page_flush(page) 45 sync_icache_dcache(struct page *page) argument 66 flush_dcache_page(struct page *page) argument [all...] |
/linux-master/net/9p/ |
H A D | trans_common.h | 7 void p9_release_pages(struct page **pages, int nr_pages);
|
/linux-master/include/net/ |
H A D | netmem.h | 15 * A netmem_ref currently is always a reference to a struct page. This 22 /* This conversion fails (returns NULL) if the netmem_ref is not struct page 25 * Currently struct page is the only possible netmem, and this helper never 28 static inline struct page *netmem_to_page(netmem_ref netmem) 30 return (__force struct page *)netmem; 33 /* Converting from page to netmem is always safe, because a page can always be 36 static inline netmem_ref page_to_netmem(struct page *page) argument 38 return (__force netmem_ref)page; [all...] |
/linux-master/include/linux/ |
H A D | bootmem_info.h | 9 * Types for free bootmem stored in page->lru.next. These have to be in 23 void get_page_bootmem(unsigned long info, struct page *page, 25 void put_page_bootmem(struct page *page); 32 static inline void free_bootmem_page(struct page *page) argument 34 unsigned long magic = page->index; 40 VM_BUG_ON_PAGE(page_ref_count(page) != 2, page); local 52 put_page_bootmem(struct page *page) argument 56 get_page_bootmem(unsigned long info, struct page *page, unsigned long type) argument 61 free_bootmem_page(struct page *page) argument [all...] |
H A D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned short order); 12 extern void __set_page_owner(struct page *page, 14 extern void __split_page_owner(struct page *page, int old_order, 17 extern void __set_page_owner_migrate_reason(struct page *page, int reason); 18 extern void __dump_page_owner(const struct page *page); 22 reset_page_owner(struct page *page, unsigned short order) argument 28 set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) argument 35 split_page_owner(struct page *page, int old_order, int new_order) argument 46 set_page_owner_migrate_reason(struct page *page, int reason) argument 51 dump_page_owner(const struct page *page) argument 57 reset_page_owner(struct page *page, unsigned short order) argument 60 set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) argument 64 split_page_owner(struct page *page, int old_order, int new_order) argument 71 set_page_owner_migrate_reason(struct page *page, int reason) argument 74 dump_page_owner(const struct page *page) argument [all...] |
H A D | page_ref.h | 7 #include <linux/page-flags.h> 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *pag 41 __page_ref_set(struct page *page, int v) argument 44 __page_ref_mod(struct page *page, int v) argument 47 __page_ref_mod_and_test(struct page *page, int v, int ret) argument 50 __page_ref_mod_and_return(struct page *page, int v, int ret) argument 53 __page_ref_mod_unless(struct page *page, int v, int u) argument 56 __page_ref_freeze(struct page *page, int v, int ret) argument 59 __page_ref_unfreeze(struct page *page, int v) argument 65 page_ref_count(const struct page *page) argument 92 page_count(const struct page *page) argument 97 set_page_count(struct page *page, int v) argument 113 init_page_count(struct page *page) argument 118 page_ref_add(struct page *page, int nr) argument 130 page_ref_sub(struct page *page, int nr) argument 142 page_ref_sub_return(struct page *page, int nr) argument 156 page_ref_inc(struct page *page) argument 168 page_ref_dec(struct page *page) argument 180 page_ref_sub_and_test(struct page *page, int nr) argument 194 page_ref_inc_return(struct page *page) argument 208 page_ref_dec_and_test(struct page *page) argument 222 page_ref_dec_return(struct page *page) argument 236 page_ref_add_unless(struct page *page, int nr, int u) argument 316 page_ref_freeze(struct page *page, int count) argument 330 page_ref_unfreeze(struct page *page, int count) argument 332 VM_BUG_ON_PAGE(page_count(page) != 0, page); local [all...] |
H A D | balloon_compaction.h | 7 * Balloon page migration makes use of the general non-lru movable page 10 * page->private is used to reference the responsible balloon device. 11 * page->mapping is used in context of non-lru page migration to reference 12 * the address space operations for page isolation/migration/compaction. 14 * As the page isolation scanning step a compaction thread does is a lockless 15 * procedure (from a page standpoint), it might bring some racy situations while 16 * performing balloon page compaction. In order to sort out these racy scenarios 17 * and safely perform balloon's page compactio 91 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) argument 108 balloon_page_delete(struct page *page) argument 125 balloon_page_device(struct page *page) argument 137 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) argument 144 balloon_page_delete(struct page *page) argument 164 balloon_page_push(struct list_head *pages, struct page *page) argument 178 struct page *page = list_first_entry_or_null(pages, struct page, lru); local [all...] |
H A D | highmem-internal.h | 10 void *__kmap_local_page_prot(struct page *page, pgprot_t prot); 35 void *kmap_high(struct page *page); 36 void kunmap_high(struct page *page); 38 struct page *__kmap_to_page(void *addr); 40 static inline void *kmap(struct page *page) argument 45 if (!PageHighMem(page)) 53 kunmap(struct page *page) argument 71 kmap_local_page(struct page *page) argument 78 struct page *page = folio_page(folio, offset / PAGE_SIZE); local 82 kmap_local_page_prot(struct page *page, pgprot_t prot) argument 97 kmap_atomic_prot(struct page *page, pgprot_t prot) argument 108 kmap_atomic(struct page *page) argument 167 kmap(struct page *page) argument 173 kunmap_high(struct page *page) argument 176 kunmap(struct page *page) argument 183 kmap_local_page(struct page *page) argument 193 kmap_local_page_prot(struct page *page, pgprot_t prot) argument 210 kmap_atomic(struct page *page) argument 220 kmap_atomic_prot(struct page *page, pgprot_t prot) argument [all...] |
H A D | page-isolation.h | 10 static inline bool is_migrate_isolate_page(struct page *page) argument 12 return get_pageblock_migratetype(page) == MIGRATE_ISOLATE; 23 static inline bool is_migrate_isolate_page(struct page *page) argument 36 void set_pageblock_migratetype(struct page *page, int migratetype); 37 int move_freepages_block(struct zone *zone, struct page *page,
|
H A D | pageblock-flags.h | 35 /* Huge page sizes are variable */ 62 struct page; 64 unsigned long get_pfnblock_flags_mask(const struct page *page, 68 void set_pfnblock_flags_mask(struct page *page, 75 #define get_pageblock_skip(page) \ 76 get_pfnblock_flags_mask(page, page_to_pfn(page), \ 78 #define clear_pageblock_skip(page) \ 86 get_pageblock_skip(struct page *page) argument 90 clear_pageblock_skip(struct page *page) argument 93 set_pageblock_skip(struct page *page) argument [all...] |
/linux-master/arch/s390/mm/ |
H A D | page-states.c | 5 * Guest page hinting for unused pages. 11 #include <asm/page-states.h> 13 #include <asm/page.h> 17 void arch_free_page(struct page *page, int order) argument 21 __set_page_unused(page_to_virt(page), 1UL << order); 24 void arch_alloc_page(struct page *page, int order) argument 29 __set_page_stable_dat(page_to_virt(page), 1UL << order); 31 __set_page_stable_nodat(page_to_virt(page), [all...] |
/linux-master/mm/ |
H A D | folio-compat.c | 13 struct address_space *page_mapping(struct page *page) argument 15 return folio_mapping(page_folio(page)); 19 void unlock_page(struct page *page) argument 21 return folio_unlock(page_folio(page)); 25 void end_page_writeback(struct page *page) argument 27 return folio_end_writeback(page_folio(page)); 31 void wait_on_page_writeback(struct page *pag argument 37 wait_for_stable_page(struct page *page) argument 43 mark_page_accessed(struct page *page) argument 49 set_page_writeback(struct page *page) argument 55 set_page_dirty(struct page *page) argument 61 __set_page_dirty_nobuffers(struct page *page) argument 67 clear_page_dirty_for_io(struct page *page) argument 73 redirty_page_for_writepage(struct writeback_control *wbc, struct page *page) argument 80 add_to_page_cache_lru(struct page *page, struct address_space *mapping, pgoff_t index, gfp_t gfp) argument 108 isolate_lru_page(struct page *page) argument 115 putback_lru_page(struct page *page) argument [all...] |
H A D | debug_page_ref.c | 8 void __page_ref_set(struct page *page, int v) argument 10 trace_page_ref_set(page, v); 15 void __page_ref_mod(struct page *page, int v) argument 17 trace_page_ref_mod(page, v); 22 void __page_ref_mod_and_test(struct page *page, int v, int ret) argument 24 trace_page_ref_mod_and_test(page, v, ret); 29 void __page_ref_mod_and_return(struct page *pag argument 36 __page_ref_mod_unless(struct page *page, int v, int u) argument 43 __page_ref_freeze(struct page *page, int v, int ret) argument 50 __page_ref_unfreeze(struct page *page, int v) argument [all...] |
H A D | bootmem_info.c | 17 void get_page_bootmem(unsigned long info, struct page *page, unsigned long type) argument 19 page->index = type; 20 SetPagePrivate(page); 21 set_page_private(page, info); 22 page_ref_inc(page); 25 void put_page_bootmem(struct page *page) argument 27 unsigned long type = page->index; 32 if (page_ref_dec_return(page) 47 struct page *page, *memmap; local 82 struct page *page, *memmap; local 106 struct page *page; local [all...] |
H A D | page_isolation.c | 7 #include <linux/page-isolation.h> 28 * Returns a page without holding a reference. If the caller wants to 29 * dereference that page (e.g., dumping), it has to make sure that it 33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, 36 struct page *page = pfn_to_page(start_pfn); local 37 struct zone *zone = page_zone(page); 43 if (is_migrate_cma_page(page)) { 52 return page; 56 page 147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) argument 206 unset_migratetype_isolate(struct page *page, int migratetype) argument 272 struct page *page; local 366 struct page *page = __first_valid_page(pfn, boundary_pfn - pfn); local 533 struct page *page; local 587 struct page *page; local 611 struct page *page; local 659 struct page *page; local [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | string_32.h | 13 #include <asm/page.h>
|
H A D | page.h | 5 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
/linux-master/arch/arm/include/asm/xen/ |
H A D | page.h | 1 #include <xen/arm/page.h>
|
/linux-master/arch/csky/abiv2/inc/abi/ |
H A D | page.h | 4 struct page *page) 10 struct page *page) 3 clear_user_page(void *addr, unsigned long vaddr, struct page *page) argument 9 copy_user_page(void *to, void *from, unsigned long vaddr, struct page *page) argument
|
/linux-master/arch/x86/include/asm/ |
H A D | agp.h | 13 * page. This avoids data corruption on some CPUs. 16 #define map_page_into_agp(page) set_pages_uc(page, 1) 17 #define unmap_page_from_agp(page) set_pages_wb(page, 1) 21 * need to be called for each cacheline of the whole page so it may 22 * not be worth it. Would need a page for it.
|
/linux-master/arch/csky/abiv1/inc/abi/ |
H A D | page.h | 5 extern void flush_dcache_page(struct page *page); 14 struct page *page) 18 flush_dcache_page(page); 22 struct page *page) 26 flush_dcache_page(page); 13 clear_user_page(void *addr, unsigned long vaddr, struct page *page) argument 21 copy_user_page(void *to, void *from, unsigned long vaddr, struct page *page) argument
|
/linux-master/include/asm-generic/ |
H A D | agp.h | 7 #define map_page_into_agp(page) do {} while (0) 8 #define unmap_page_from_agp(page) do {} while (0)
|
/linux-master/arch/arm64/include/asm/ |
H A D | set_memory.h | 13 int set_direct_map_invalid_noflush(struct page *page); 14 int set_direct_map_default_noflush(struct page *page); 15 bool kernel_page_present(struct page *page);
|
/linux-master/include/net/page_pool/ |
H A D | helpers.h | 11 * The page_pool allocator is optimized for recycling page or page fragment used 15 * which allocate memory with or without page splitting depending on the 19 * always smaller than half a page, it can use one of the more specific API 22 * 1. page_pool_alloc_pages(): allocate memory without page splitting when 23 * driver knows that the memory it need is always bigger than half of the page 24 * allocated from page pool. There is no cache line dirtying for 'struct page' 25 * when a page is recycled back to the page poo 120 struct page *page; local 170 struct page *page; local 231 page_pool_fragment_page(struct page *page, long nr) argument 236 page_pool_unref_page(struct page *page, long nr) argument 280 page_pool_ref_page(struct page *page) argument 285 page_pool_is_last_ref(struct page *page) argument 304 page_pool_put_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) argument 329 page_pool_put_full_page(struct page_pool *pool, struct page *page, bool allow_direct) argument 343 page_pool_recycle_direct(struct page_pool *pool, struct page *page) argument 373 page_pool_get_dma_addr(struct page *page) argument 383 page_pool_set_dma_addr(struct page *page, dma_addr_t addr) argument [all...] |