Searched refs:page (Results 26 - 50 of 3144) sorted by relevance

1234567891011>>

/linux-master/tools/virtio/linux/
H A Dkmsan.h7 inline void kmsan_handle_dma(struct page *page, size_t offset, size_t size, argument
/linux-master/arch/x86/include/asm/
H A Dset_memory.h6 #include <asm/page.h>
32 * operation on other mappings of the same physical page
33 * - Restore default attributes when a page is freed
58 int set_pages_array_uc(struct page **pages, int addrinarray);
59 int set_pages_array_wc(struct page **pages, int addrinarray);
60 int set_pages_array_wb(struct page **pages, int addrinarray);
64 * are provided that work on a "struct page".
66 * memory that the struct page represents, and internally just
77 * get a struct page* that the old API required.
82 int set_pages_uc(struct page *pag
[all...]
/linux-master/include/linux/
H A Dpage_idle.h6 #include <linux/page-flags.h>
13 * If there is not enough space to store Idle and Young bits in page flags, use
14 * page ext flags instead.
18 struct page_ext *page_ext = page_ext_get(&folio->page);
32 struct page_ext *page_ext = page_ext_get(&folio->page);
43 struct page_ext *page_ext = page_ext_get(&folio->page);
57 struct page_ext *page_ext = page_ext_get(&folio->page);
71 struct page_ext *page_ext = page_ext_get(&folio->page);
82 struct page_ext *page_ext = page_ext_get(&folio->page);
123 static inline bool page_is_young(struct page *pag argument
128 set_page_young(struct page *page) argument
133 test_and_clear_page_young(struct page *page) argument
138 page_is_idle(struct page *page) argument
143 set_page_idle(struct page *page) argument
[all...]
H A Dpage-flags.h3 * Macros for manipulating and testing page->flags
18 * Various page->flags bits:
20 * PG_reserved is set for special pages. The "struct page" of such a page
25 * - Pages reserved or allocated early during boot (before the page allocator
27 * initial vmemmap, initial page tables, crashkernel, elfcorehdr, and much
29 * be given to the page allocator.
32 * - The zero page(s)
33 * - Pages not added to the page allocator when onlining a section because
49 * Consequently, PG_reserved for a page mappe
207 page_fixed_fake_head(const struct page *page) argument
233 page_fixed_fake_head(const struct page *page) argument
239 page_is_fake_head(const struct page *page) argument
244 _compound_head(const struct page *page) argument
283 PageTail(const struct page *page) argument
288 PageCompound(const struct page *page) argument
295 PagePoisoned(const struct page *page) argument
303 page_init_poison(struct page *page, size_t size) argument
311 const struct page *page = &folio->page; local
313 VM_BUG_ON_PGFLAGS(PageTail(page), page); local
320 struct page *page = &folio->page; local
322 VM_BUG_ON_PGFLAGS(PageTail(page), page); local
587 PageSwapCache(const struct page *page) argument
691 PageMappingFlags(const struct page *page) argument
701 PageAnon(const struct page *page) argument
712 __PageMovable(const struct page *page) argument
731 PageKsm(const struct page *page) argument
787 PageUptodate(const struct page *page) argument
809 __SetPageUptodate(struct page *page) argument
814 SetPageUptodate(struct page *page) argument
834 PageHead(const struct page *page) argument
855 set_compound_head(struct page *page, struct page *head) argument
860 clear_compound_head(struct page *page) argument
866 ClearPageCompound(struct page *page) argument
887 PageTransHuge(const struct page *page) argument
889 VM_BUG_ON_PAGE(PageTail(page), page); local
898 PageTransCompound(const struct page *page) argument
908 PageTransTail(const struct page *page) argument
961 page_has_type(const struct page *page) argument
1058 PageHuge(const struct page *page) argument
1068 is_page_hwpoison(struct page *page) argument
1079 PageAnonExclusive(const struct page *page) argument
1082 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local
1086 SetPageAnonExclusive(struct page *page) argument
1089 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local
1093 ClearPageAnonExclusive(struct page *page) argument
1096 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local
1100 __ClearPageAnonExclusive(struct page *page) argument
1103 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local
1152 page_has_private(const struct page *page) argument
[all...]
H A Dcacheflush.h22 struct page *page, unsigned int nr)
27 #define flush_icache_page(vma, page) flush_icache_pages(vma, page, 1)
21 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument
H A Dkmsan.h17 struct page;
51 * @page: struct page to free.
52 * @order: order of @page.
57 bool __init __must_check kmsan_memblock_free_pages(struct page *page,
62 * @page: struct page pointer returned by alloc_pages().
63 * @order: order of allocated struct page.
66 * KMSAN marks 1<<@order pages starting at @page a
243 kmsan_memblock_free_pages(struct page *page, unsigned int order) argument
257 kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags) argument
262 kmsan_free_page(struct page *page, unsigned int order) argument
314 kmsan_handle_dma(struct page *page, size_t offset, size_t size, enum dma_data_direction dir) argument
[all...]
/linux-master/mm/
H A Dballoon_compaction.c15 struct page *page)
18 * Block others from accessing the 'page' when we get around to
20 * holding a reference to the 'page' at this point. If we are not, then
23 BUG_ON(!trylock_page(page));
24 balloon_page_insert(b_dev_info, page);
25 unlock_page(page);
30 * balloon_page_list_enqueue() - inserts a list of pages into the balloon page
32 * @b_dev_info: balloon device descriptor where we will insert a new page to
43 struct page *pag local
14 balloon_page_enqueue_one(struct balloon_dev_info *b_dev_info, struct page *page) argument
79 struct page *page, *tmp; local
126 struct page *page = alloc_page(balloon_mapping_gfp_mask() | local
146 balloon_page_enqueue(struct balloon_dev_info *b_dev_info, struct page *page) argument
206 balloon_page_isolate(struct page *page, isolate_mode_t mode) argument
220 balloon_page_putback(struct page *page) argument
232 balloon_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) argument
[all...]
H A Dpage_poison.c22 static void poison_page(struct page *page) argument
24 void *addr = kmap_local_page(page);
26 /* KASAN still think the page is in-use, so skip it. */
33 void __kernel_poison_pages(struct page *page, int n) argument
38 poison_page(page + i);
48 static void check_poison_mem(struct page *page, unsigned char *mem, size_t bytes) argument
73 dump_page(page, "pageallo
76 unpoison_page(struct page *page) argument
92 __kernel_unpoison_pages(struct page *page, int n) argument
101 __kernel_map_pages(struct page *page, int numpages, int enable) argument
[all...]
/linux-master/arch/arm/include/asm/
H A Dpage.h3 * arch/arm/include/asm/page.h
10 /* PAGE_SHIFT determines the page size */
19 #include <asm/page-nommu.h>
30 * page-based copying and clearing for user space for the particular
107 struct page;
111 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr);
112 void (*cpu_copy_user_highpage)(struct page *to, struct page *from,
116 void fa_copy_user_highpage(struct page *t
[all...]
H A Dsparsemem.h5 #include <asm/page.h>
H A Dhugetlb.h14 #include <asm/page.h>
18 static inline void arch_clear_hugepage_flags(struct page *page) argument
20 clear_bit(PG_dcache_clean, &page->flags);
/linux-master/arch/sparc/include/asm/
H A Dsparsemem.h7 #include <asm/page.h>
/linux-master/arch/arm64/include/asm/xen/
H A Dpage.h1 #include <xen/arm/page.h>
/linux-master/arch/arm64/include/asm/
H A Dpage.h3 * Based on arch/arm/include/asm/page.h
11 #include <asm/page-def.h>
19 struct page;
25 void copy_user_highpage(struct page *to, struct page *from,
29 void copy_highpage(struct page *to, struct page *from);
36 void tag_clear_highpage(struct page *to);
39 #define clear_user_page(page, vaddr, pg) clear_page(page)
[all...]
/linux-master/include/trace/events/
H A Dpage_ref.h15 TP_PROTO(struct page *page, int v),
17 TP_ARGS(page, v),
30 __entry->pfn = page_to_pfn(page);
31 __entry->flags = page->flags;
32 __entry->count = page_ref_count(page);
33 __entry->mapcount = page_mapcount(page);
34 __entry->mapping = page->mapping;
35 __entry->mt = get_pageblock_migratetype(page);
49 TP_PROTO(struct page *pag
[all...]
H A Dcma.h13 TP_PROTO(const char *name, unsigned long pfn, const struct page *page,
16 TP_ARGS(name, pfn, page, count),
21 __field(const struct page *, page)
28 __entry->page = page;
32 TP_printk("name=%s pfn=0x%lx page=%p count=%lu",
35 __entry->page,
65 TP_PROTO(const char *name, unsigned long pfn, const struct page *pag
[all...]
/linux-master/fs/erofs/
H A Dcompress.h13 struct page **in, **out;
26 struct page **pagepool);
30 /* some special page->private (unsigned long, see below) */
35 * For all pages in a pcluster, page->private should be one of
36 * Type Last 2bits page->private
37 * short-lived page 00 Z_EROFS_SHORTLIVED_PAGE
38 * preallocated page (tryalloc) 00 Z_EROFS_PREALLOCATED_PAGE
39 * cached/managed page 00 pointer to z_erofs_pcluster
40 * online page (file-backed, 01/10/11 sub-index << 2 | count
43 * page
59 z_erofs_is_shortlived_page(struct page *page) argument
68 z_erofs_put_shortlivedpage(struct page **pagepool, struct page *page) argument
[all...]
/linux-master/sound/pci/trident/
H A Dtrident_memory.c7 * Trident 4DWave-NX memory page allocation (TLB area)
19 /* page arguments of these two macros are Trident page (4096 bytes), not like
22 #define __set_tlb_bus(trident,page,addr) \
23 (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1))
24 #define __tlb_to_addr(trident,page) \
25 (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1))
28 /* page size == SNDRV_TRIDENT_PAGE_SIZE */
29 #define ALIGN_PAGE_SIZE PAGE_SIZE /* minimum page size for allocation */
31 /* fill TLB entrie(s) corresponding to page wit
51 set_tlb_bus(struct snd_trident *trident, int page, dma_addr_t addr) argument
58 set_silent_tlb(struct snd_trident *trident, int page) argument
80 set_tlb_bus(struct snd_trident *trident, int page, dma_addr_t addr) argument
90 set_silent_tlb(struct snd_trident *trident, int page) argument
111 int page, psize; local
163 int idx, page; local
207 int page; local
264 int page; local
[all...]
/linux-master/fs/squashfs/
H A Dfile_direct.c21 /* Read separately compressed datablock directly into page cache */
22 int squashfs_readpage_block(struct page *target_page, u64 block, int bsize,
34 struct page **page; local
43 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL);
44 if (page == NULL)
49 page[i] = (n == target_page->index) ? target_page :
52 if (page[i] == NULL)
55 if (PageUptodate(page[i])) {
56 unlock_page(page[
[all...]
/linux-master/arch/openrisc/mm/
H A Dcache.c19 static __always_inline void cache_loop(struct page *page, const unsigned int reg) argument
21 unsigned long paddr = page_to_pfn(page) << PAGE_SHIFT;
30 void local_dcache_page_flush(struct page *page) argument
32 cache_loop(page, SPR_DCBFR);
36 void local_icache_page_inv(struct page *page) argument
38 cache_loop(page, SPR_ICBIR);
/linux-master/drivers/infiniband/hw/mlx5/
H A Ddoorbell.c51 struct mlx5_ib_user_db_page *page; local
56 list_for_each_entry(page, &context->db_page_list, list)
57 if ((current->mm == page->mm) &&
58 (page->user_virt == (virt & PAGE_MASK)))
61 page = kmalloc(sizeof(*page), GFP_KERNEL);
62 if (!page) {
67 page->user_virt = (virt & PAGE_MASK);
68 page->refcnt = 0;
69 page
[all...]
/linux-master/net/ceph/
H A Dpagelist.c32 struct page *page = list_entry(pl->head.prev, struct page, lru); local
33 kunmap(page);
44 struct page *page = list_first_entry(&pl->head, struct page, local
46 list_del(&page->lru);
47 __free_page(page);
56 struct page *pag local
110 struct page *page = __page_cache_alloc(GFP_NOFS); local
124 struct page *page = list_first_entry(&pl->free_list, local
153 struct page *page; local
[all...]
/linux-master/arch/sh/include/asm/
H A Dcacheflush.h13 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page
47 static inline void flush_dcache_page(struct page *page) argument
49 flush_dcache_folio(page_folio(page));
54 void flush_icache_pages(struct vm_area_struct *vma, struct page *page,
65 extern void __flush_anon_page(struct page *page, unsigned long);
68 struct page *page, unsigne
67 flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument
[all...]
H A Dtlbflush.h10 * - flush_tlb_page(vma, vmaddr) flushes one page
20 unsigned long page);
23 extern void local_flush_tlb_one(unsigned long asid, unsigned long page);
33 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
35 extern void flush_tlb_one(unsigned long asid, unsigned long page);
41 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
42 #define flush_tlb_one(asid, page) local_flush_tlb_one(asid, page)
/linux-master/fs/sysv/
H A Ddir.c31 static void dir_commit_chunk(struct page *page, loff_t pos, unsigned len) argument
33 struct address_space *mapping = page->mapping;
36 block_write_end(NULL, mapping, pos, len, len, page, NULL);
41 unlock_page(page);
61 static void *dir_get_page(struct inode *dir, unsigned long n, struct page **p)
64 struct page *page = read_mapping_page(mapping, n, NULL); local
65 if (IS_ERR(page))
66 return ERR_CAST(page);
90 struct page *page; local
146 struct page *page = NULL; local
189 struct page *page = NULL; local
238 sysv_delete_entry(struct sysv_dir_entry *de, struct page *page) argument
259 struct page *page = grab_cache_page(inode->i_mapping, 0); local
295 struct page *page = NULL; local
334 sysv_set_link(struct sysv_dir_entry *de, struct page *page, struct inode *inode) argument
373 struct page *page; local
[all...]

Completed in 410 milliseconds

1234567891011>>