Searched refs:page (Results 1 - 25 of 3144) sorted by relevance

1234567891011>>

/linux-master/tools/virtio/linux/
H A Dmm_types.h2 struct page page; member in struct:folio
/linux-master/arch/openrisc/include/asm/
H A Dcacheflush.h24 extern void local_dcache_page_flush(struct page *page);
25 extern void local_icache_page_inv(struct page *page);
33 #define dcache_page_flush(page) local_dcache_page_flush(page)
34 #define icache_page_inv(page) local_icache_page_inv(page)
36 #define dcache_page_flush(page) local_dcache_page_flush(page)
45 sync_icache_dcache(struct page *page) argument
66 flush_dcache_page(struct page *page) argument
[all...]
/linux-master/net/9p/
H A Dtrans_common.h7 void p9_release_pages(struct page **pages, int nr_pages);
/linux-master/include/net/
H A Dnetmem.h15 * A netmem_ref currently is always a reference to a struct page. This
22 /* This conversion fails (returns NULL) if the netmem_ref is not struct page
25 * Currently struct page is the only possible netmem, and this helper never
28 static inline struct page *netmem_to_page(netmem_ref netmem)
30 return (__force struct page *)netmem;
33 /* Converting from page to netmem is always safe, because a page can always be
36 static inline netmem_ref page_to_netmem(struct page *page) argument
38 return (__force netmem_ref)page;
[all...]
/linux-master/include/linux/
H A Dbootmem_info.h9 * Types for free bootmem stored in page->lru.next. These have to be in
23 void get_page_bootmem(unsigned long info, struct page *page,
25 void put_page_bootmem(struct page *page);
32 static inline void free_bootmem_page(struct page *page) argument
34 unsigned long magic = page->index;
40 VM_BUG_ON_PAGE(page_ref_count(page) != 2, page); local
52 put_page_bootmem(struct page *page) argument
56 get_page_bootmem(unsigned long info, struct page *page, unsigned long type) argument
61 free_bootmem_page(struct page *page) argument
[all...]
H A Dpage_owner.h11 extern void __reset_page_owner(struct page *page, unsigned short order);
12 extern void __set_page_owner(struct page *page,
14 extern void __split_page_owner(struct page *page, int old_order,
17 extern void __set_page_owner_migrate_reason(struct page *page, int reason);
18 extern void __dump_page_owner(const struct page *page);
22 reset_page_owner(struct page *page, unsigned short order) argument
28 set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) argument
35 split_page_owner(struct page *page, int old_order, int new_order) argument
46 set_page_owner_migrate_reason(struct page *page, int reason) argument
51 dump_page_owner(const struct page *page) argument
57 reset_page_owner(struct page *page, unsigned short order) argument
60 set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) argument
64 split_page_owner(struct page *page, int old_order, int new_order) argument
71 set_page_owner_migrate_reason(struct page *page, int reason) argument
74 dump_page_owner(const struct page *page) argument
[all...]
H A Dpage_ref.h7 #include <linux/page-flags.h>
29 extern void __page_ref_set(struct page *page, int v);
30 extern void __page_ref_mod(struct page *page, int v);
31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret);
32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret);
33 extern void __page_ref_mod_unless(struct page *pag
41 __page_ref_set(struct page *page, int v) argument
44 __page_ref_mod(struct page *page, int v) argument
47 __page_ref_mod_and_test(struct page *page, int v, int ret) argument
50 __page_ref_mod_and_return(struct page *page, int v, int ret) argument
53 __page_ref_mod_unless(struct page *page, int v, int u) argument
56 __page_ref_freeze(struct page *page, int v, int ret) argument
59 __page_ref_unfreeze(struct page *page, int v) argument
65 page_ref_count(const struct page *page) argument
92 page_count(const struct page *page) argument
97 set_page_count(struct page *page, int v) argument
113 init_page_count(struct page *page) argument
118 page_ref_add(struct page *page, int nr) argument
130 page_ref_sub(struct page *page, int nr) argument
142 page_ref_sub_return(struct page *page, int nr) argument
156 page_ref_inc(struct page *page) argument
168 page_ref_dec(struct page *page) argument
180 page_ref_sub_and_test(struct page *page, int nr) argument
194 page_ref_inc_return(struct page *page) argument
208 page_ref_dec_and_test(struct page *page) argument
222 page_ref_dec_return(struct page *page) argument
236 page_ref_add_unless(struct page *page, int nr, int u) argument
316 page_ref_freeze(struct page *page, int count) argument
330 page_ref_unfreeze(struct page *page, int count) argument
332 VM_BUG_ON_PAGE(page_count(page) != 0, page); local
[all...]
H A Dballoon_compaction.h7 * Balloon page migration makes use of the general non-lru movable page
10 * page->private is used to reference the responsible balloon device.
11 * page->mapping is used in context of non-lru page migration to reference
12 * the address space operations for page isolation/migration/compaction.
14 * As the page isolation scanning step a compaction thread does is a lockless
15 * procedure (from a page standpoint), it might bring some racy situations while
16 * performing balloon page compaction. In order to sort out these racy scenarios
17 * and safely perform balloon's page compactio
91 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) argument
108 balloon_page_delete(struct page *page) argument
125 balloon_page_device(struct page *page) argument
137 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) argument
144 balloon_page_delete(struct page *page) argument
164 balloon_page_push(struct list_head *pages, struct page *page) argument
178 struct page *page = list_first_entry_or_null(pages, struct page, lru); local
[all...]
H A Dhighmem-internal.h10 void *__kmap_local_page_prot(struct page *page, pgprot_t prot);
35 void *kmap_high(struct page *page);
36 void kunmap_high(struct page *page);
38 struct page *__kmap_to_page(void *addr);
40 static inline void *kmap(struct page *page) argument
45 if (!PageHighMem(page))
53 kunmap(struct page *page) argument
71 kmap_local_page(struct page *page) argument
78 struct page *page = folio_page(folio, offset / PAGE_SIZE); local
82 kmap_local_page_prot(struct page *page, pgprot_t prot) argument
97 kmap_atomic_prot(struct page *page, pgprot_t prot) argument
108 kmap_atomic(struct page *page) argument
167 kmap(struct page *page) argument
173 kunmap_high(struct page *page) argument
176 kunmap(struct page *page) argument
183 kmap_local_page(struct page *page) argument
193 kmap_local_page_prot(struct page *page, pgprot_t prot) argument
210 kmap_atomic(struct page *page) argument
220 kmap_atomic_prot(struct page *page, pgprot_t prot) argument
[all...]
H A Dpage-isolation.h10 static inline bool is_migrate_isolate_page(struct page *page) argument
12 return get_pageblock_migratetype(page) == MIGRATE_ISOLATE;
23 static inline bool is_migrate_isolate_page(struct page *page) argument
36 void set_pageblock_migratetype(struct page *page, int migratetype);
37 int move_freepages_block(struct zone *zone, struct page *page,
H A Dpageblock-flags.h35 /* Huge page sizes are variable */
62 struct page;
64 unsigned long get_pfnblock_flags_mask(const struct page *page,
68 void set_pfnblock_flags_mask(struct page *page,
75 #define get_pageblock_skip(page) \
76 get_pfnblock_flags_mask(page, page_to_pfn(page), \
78 #define clear_pageblock_skip(page) \
86 get_pageblock_skip(struct page *page) argument
90 clear_pageblock_skip(struct page *page) argument
93 set_pageblock_skip(struct page *page) argument
[all...]
/linux-master/arch/s390/mm/
H A Dpage-states.c5 * Guest page hinting for unused pages.
11 #include <asm/page-states.h>
13 #include <asm/page.h>
17 void arch_free_page(struct page *page, int order) argument
21 __set_page_unused(page_to_virt(page), 1UL << order);
24 void arch_alloc_page(struct page *page, int order) argument
29 __set_page_stable_dat(page_to_virt(page), 1UL << order);
31 __set_page_stable_nodat(page_to_virt(page),
[all...]
/linux-master/mm/
H A Dfolio-compat.c13 struct address_space *page_mapping(struct page *page) argument
15 return folio_mapping(page_folio(page));
19 void unlock_page(struct page *page) argument
21 return folio_unlock(page_folio(page));
25 void end_page_writeback(struct page *page) argument
27 return folio_end_writeback(page_folio(page));
31 void wait_on_page_writeback(struct page *pag argument
37 wait_for_stable_page(struct page *page) argument
43 mark_page_accessed(struct page *page) argument
49 set_page_writeback(struct page *page) argument
55 set_page_dirty(struct page *page) argument
61 __set_page_dirty_nobuffers(struct page *page) argument
67 clear_page_dirty_for_io(struct page *page) argument
73 redirty_page_for_writepage(struct writeback_control *wbc, struct page *page) argument
80 add_to_page_cache_lru(struct page *page, struct address_space *mapping, pgoff_t index, gfp_t gfp) argument
108 isolate_lru_page(struct page *page) argument
115 putback_lru_page(struct page *page) argument
[all...]
H A Ddebug_page_ref.c8 void __page_ref_set(struct page *page, int v) argument
10 trace_page_ref_set(page, v);
15 void __page_ref_mod(struct page *page, int v) argument
17 trace_page_ref_mod(page, v);
22 void __page_ref_mod_and_test(struct page *page, int v, int ret) argument
24 trace_page_ref_mod_and_test(page, v, ret);
29 void __page_ref_mod_and_return(struct page *pag argument
36 __page_ref_mod_unless(struct page *page, int v, int u) argument
43 __page_ref_freeze(struct page *page, int v, int ret) argument
50 __page_ref_unfreeze(struct page *page, int v) argument
[all...]
H A Dbootmem_info.c17 void get_page_bootmem(unsigned long info, struct page *page, unsigned long type) argument
19 page->index = type;
20 SetPagePrivate(page);
21 set_page_private(page, info);
22 page_ref_inc(page);
25 void put_page_bootmem(struct page *page) argument
27 unsigned long type = page->index;
32 if (page_ref_dec_return(page)
47 struct page *page, *memmap; local
82 struct page *page, *memmap; local
106 struct page *page; local
[all...]
H A Dpage_isolation.c7 #include <linux/page-isolation.h>
28 * Returns a page without holding a reference. If the caller wants to
29 * dereference that page (e.g., dumping), it has to make sure that it
33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn,
36 struct page *page = pfn_to_page(start_pfn); local
37 struct zone *zone = page_zone(page);
43 if (is_migrate_cma_page(page)) {
52 return page;
56 page
147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) argument
206 unset_migratetype_isolate(struct page *page, int migratetype) argument
272 struct page *page; local
366 struct page *page = __first_valid_page(pfn, boundary_pfn - pfn); local
533 struct page *page; local
587 struct page *page; local
611 struct page *page; local
659 struct page *page; local
[all...]
/linux-master/arch/sparc/include/asm/
H A Dstring_32.h13 #include <asm/page.h>
H A Dpage.h5 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
/linux-master/arch/arm/include/asm/xen/
H A Dpage.h1 #include <xen/arm/page.h>
/linux-master/arch/csky/abiv2/inc/abi/
H A Dpage.h4 struct page *page)
10 struct page *page)
3 clear_user_page(void *addr, unsigned long vaddr, struct page *page) argument
9 copy_user_page(void *to, void *from, unsigned long vaddr, struct page *page) argument
/linux-master/arch/x86/include/asm/
H A Dagp.h13 * page. This avoids data corruption on some CPUs.
16 #define map_page_into_agp(page) set_pages_uc(page, 1)
17 #define unmap_page_from_agp(page) set_pages_wb(page, 1)
21 * need to be called for each cacheline of the whole page so it may
22 * not be worth it. Would need a page for it.
/linux-master/arch/csky/abiv1/inc/abi/
H A Dpage.h5 extern void flush_dcache_page(struct page *page);
14 struct page *page)
18 flush_dcache_page(page);
22 struct page *page)
26 flush_dcache_page(page);
13 clear_user_page(void *addr, unsigned long vaddr, struct page *page) argument
21 copy_user_page(void *to, void *from, unsigned long vaddr, struct page *page) argument
/linux-master/include/asm-generic/
H A Dagp.h7 #define map_page_into_agp(page) do {} while (0)
8 #define unmap_page_from_agp(page) do {} while (0)
/linux-master/arch/arm64/include/asm/
H A Dset_memory.h13 int set_direct_map_invalid_noflush(struct page *page);
14 int set_direct_map_default_noflush(struct page *page);
15 bool kernel_page_present(struct page *page);
/linux-master/include/net/page_pool/
H A Dhelpers.h11 * The page_pool allocator is optimized for recycling page or page fragment used
15 * which allocate memory with or without page splitting depending on the
19 * always smaller than half a page, it can use one of the more specific API
22 * 1. page_pool_alloc_pages(): allocate memory without page splitting when
23 * driver knows that the memory it need is always bigger than half of the page
24 * allocated from page pool. There is no cache line dirtying for 'struct page'
25 * when a page is recycled back to the page poo
120 struct page *page; local
170 struct page *page; local
231 page_pool_fragment_page(struct page *page, long nr) argument
236 page_pool_unref_page(struct page *page, long nr) argument
280 page_pool_ref_page(struct page *page) argument
285 page_pool_is_last_ref(struct page *page) argument
304 page_pool_put_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) argument
329 page_pool_put_full_page(struct page_pool *pool, struct page *page, bool allow_direct) argument
343 page_pool_recycle_direct(struct page_pool *pool, struct page *page) argument
373 page_pool_get_dma_addr(struct page *page) argument
383 page_pool_set_dma_addr(struct page *page, dma_addr_t addr) argument
[all...]

Completed in 1555 milliseconds

1234567891011>>