/linux-master/tools/virtio/linux/ |
H A D | kmsan.h | 7 inline void kmsan_handle_dma(struct page *page, size_t offset, size_t size, argument
|
/linux-master/arch/x86/include/asm/ |
H A D | set_memory.h | 6 #include <asm/page.h> 32 * operation on other mappings of the same physical page 33 * - Restore default attributes when a page is freed 58 int set_pages_array_uc(struct page **pages, int addrinarray); 59 int set_pages_array_wc(struct page **pages, int addrinarray); 60 int set_pages_array_wb(struct page **pages, int addrinarray); 64 * are provided that work on a "struct page". 66 * memory that the struct page represents, and internally just 77 * get a struct page* that the old API required. 82 int set_pages_uc(struct page *pag [all...] |
/linux-master/include/linux/ |
H A D | page_idle.h | 6 #include <linux/page-flags.h> 13 * If there is not enough space to store Idle and Young bits in page flags, use 14 * page ext flags instead. 18 struct page_ext *page_ext = page_ext_get(&folio->page); 32 struct page_ext *page_ext = page_ext_get(&folio->page); 43 struct page_ext *page_ext = page_ext_get(&folio->page); 57 struct page_ext *page_ext = page_ext_get(&folio->page); 71 struct page_ext *page_ext = page_ext_get(&folio->page); 82 struct page_ext *page_ext = page_ext_get(&folio->page); 123 static inline bool page_is_young(struct page *pag argument 128 set_page_young(struct page *page) argument 133 test_and_clear_page_young(struct page *page) argument 138 page_is_idle(struct page *page) argument 143 set_page_idle(struct page *page) argument [all...] |
H A D | page-flags.h | 3 * Macros for manipulating and testing page->flags 18 * Various page->flags bits: 20 * PG_reserved is set for special pages. The "struct page" of such a page 25 * - Pages reserved or allocated early during boot (before the page allocator 27 * initial vmemmap, initial page tables, crashkernel, elfcorehdr, and much 29 * be given to the page allocator. 32 * - The zero page(s) 33 * - Pages not added to the page allocator when onlining a section because 49 * Consequently, PG_reserved for a page mappe 207 page_fixed_fake_head(const struct page *page) argument 233 page_fixed_fake_head(const struct page *page) argument 239 page_is_fake_head(const struct page *page) argument 244 _compound_head(const struct page *page) argument 283 PageTail(const struct page *page) argument 288 PageCompound(const struct page *page) argument 295 PagePoisoned(const struct page *page) argument 303 page_init_poison(struct page *page, size_t size) argument 311 const struct page *page = &folio->page; local 313 VM_BUG_ON_PGFLAGS(PageTail(page), page); local 320 struct page *page = &folio->page; local 322 VM_BUG_ON_PGFLAGS(PageTail(page), page); local 587 PageSwapCache(const struct page *page) argument 691 PageMappingFlags(const struct page *page) argument 701 PageAnon(const struct page *page) argument 712 __PageMovable(const struct page *page) argument 731 PageKsm(const struct page *page) argument 787 PageUptodate(const struct page *page) argument 809 __SetPageUptodate(struct page *page) argument 814 SetPageUptodate(struct page *page) argument 834 PageHead(const struct page *page) argument 855 set_compound_head(struct page *page, struct page *head) argument 860 clear_compound_head(struct page *page) argument 866 ClearPageCompound(struct page *page) argument 887 PageTransHuge(const struct page *page) argument 889 VM_BUG_ON_PAGE(PageTail(page), page); local 898 PageTransCompound(const struct page *page) argument 908 PageTransTail(const struct page *page) argument 961 page_has_type(const struct page *page) argument 1058 PageHuge(const struct page *page) argument 1068 is_page_hwpoison(struct page *page) argument 1079 PageAnonExclusive(const struct page *page) argument 1082 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local 1086 SetPageAnonExclusive(struct page *page) argument 1089 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local 1093 ClearPageAnonExclusive(struct page *page) argument 1096 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local 1100 __ClearPageAnonExclusive(struct page *page) argument 1103 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); local 1152 page_has_private(const struct page *page) argument [all...] |
H A D | cacheflush.h | 22 struct page *page, unsigned int nr) 27 #define flush_icache_page(vma, page) flush_icache_pages(vma, page, 1) 21 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument
|
H A D | kmsan.h | 17 struct page; 51 * @page: struct page to free. 52 * @order: order of @page. 57 bool __init __must_check kmsan_memblock_free_pages(struct page *page, 62 * @page: struct page pointer returned by alloc_pages(). 63 * @order: order of allocated struct page. 66 * KMSAN marks 1<<@order pages starting at @page a 243 kmsan_memblock_free_pages(struct page *page, unsigned int order) argument 257 kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags) argument 262 kmsan_free_page(struct page *page, unsigned int order) argument 314 kmsan_handle_dma(struct page *page, size_t offset, size_t size, enum dma_data_direction dir) argument [all...] |
/linux-master/mm/ |
H A D | balloon_compaction.c | 15 struct page *page) 18 * Block others from accessing the 'page' when we get around to 20 * holding a reference to the 'page' at this point. If we are not, then 23 BUG_ON(!trylock_page(page)); 24 balloon_page_insert(b_dev_info, page); 25 unlock_page(page); 30 * balloon_page_list_enqueue() - inserts a list of pages into the balloon page 32 * @b_dev_info: balloon device descriptor where we will insert a new page to 43 struct page *pag local 14 balloon_page_enqueue_one(struct balloon_dev_info *b_dev_info, struct page *page) argument 79 struct page *page, *tmp; local 126 struct page *page = alloc_page(balloon_mapping_gfp_mask() | local 146 balloon_page_enqueue(struct balloon_dev_info *b_dev_info, struct page *page) argument 206 balloon_page_isolate(struct page *page, isolate_mode_t mode) argument 220 balloon_page_putback(struct page *page) argument 232 balloon_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) argument [all...] |
H A D | page_poison.c | 22 static void poison_page(struct page *page) argument 24 void *addr = kmap_local_page(page); 26 /* KASAN still think the page is in-use, so skip it. */ 33 void __kernel_poison_pages(struct page *page, int n) argument 38 poison_page(page + i); 48 static void check_poison_mem(struct page *page, unsigned char *mem, size_t bytes) argument 73 dump_page(page, "pageallo 76 unpoison_page(struct page *page) argument 92 __kernel_unpoison_pages(struct page *page, int n) argument 101 __kernel_map_pages(struct page *page, int numpages, int enable) argument [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | page.h | 3 * arch/arm/include/asm/page.h 10 /* PAGE_SHIFT determines the page size */ 19 #include <asm/page-nommu.h> 30 * page-based copying and clearing for user space for the particular 107 struct page; 111 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr); 112 void (*cpu_copy_user_highpage)(struct page *to, struct page *from, 116 void fa_copy_user_highpage(struct page *t [all...] |
H A D | sparsemem.h | 5 #include <asm/page.h>
|
H A D | hugetlb.h | 14 #include <asm/page.h> 18 static inline void arch_clear_hugepage_flags(struct page *page) argument 20 clear_bit(PG_dcache_clean, &page->flags);
|
/linux-master/arch/sparc/include/asm/ |
H A D | sparsemem.h | 7 #include <asm/page.h>
|
/linux-master/arch/arm64/include/asm/xen/ |
H A D | page.h | 1 #include <xen/arm/page.h>
|
/linux-master/arch/arm64/include/asm/ |
H A D | page.h | 3 * Based on arch/arm/include/asm/page.h 11 #include <asm/page-def.h> 19 struct page; 25 void copy_user_highpage(struct page *to, struct page *from, 29 void copy_highpage(struct page *to, struct page *from); 36 void tag_clear_highpage(struct page *to); 39 #define clear_user_page(page, vaddr, pg) clear_page(page) [all...] |
/linux-master/include/trace/events/ |
H A D | page_ref.h | 15 TP_PROTO(struct page *page, int v), 17 TP_ARGS(page, v), 30 __entry->pfn = page_to_pfn(page); 31 __entry->flags = page->flags; 32 __entry->count = page_ref_count(page); 33 __entry->mapcount = page_mapcount(page); 34 __entry->mapping = page->mapping; 35 __entry->mt = get_pageblock_migratetype(page); 49 TP_PROTO(struct page *pag [all...] |
H A D | cma.h | 13 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 16 TP_ARGS(name, pfn, page, count), 21 __field(const struct page *, page) 28 __entry->page = page; 32 TP_printk("name=%s pfn=0x%lx page=%p count=%lu", 35 __entry->page, 65 TP_PROTO(const char *name, unsigned long pfn, const struct page *pag [all...] |
/linux-master/fs/erofs/ |
H A D | compress.h | 13 struct page **in, **out; 26 struct page **pagepool); 30 /* some special page->private (unsigned long, see below) */ 35 * For all pages in a pcluster, page->private should be one of 36 * Type Last 2bits page->private 37 * short-lived page 00 Z_EROFS_SHORTLIVED_PAGE 38 * preallocated page (tryalloc) 00 Z_EROFS_PREALLOCATED_PAGE 39 * cached/managed page 00 pointer to z_erofs_pcluster 40 * online page (file-backed, 01/10/11 sub-index << 2 | count 43 * page 59 z_erofs_is_shortlived_page(struct page *page) argument 68 z_erofs_put_shortlivedpage(struct page **pagepool, struct page *page) argument [all...] |
/linux-master/sound/pci/trident/ |
H A D | trident_memory.c | 7 * Trident 4DWave-NX memory page allocation (TLB area) 19 /* page arguments of these two macros are Trident page (4096 bytes), not like 22 #define __set_tlb_bus(trident,page,addr) \ 23 (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1)) 24 #define __tlb_to_addr(trident,page) \ 25 (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1)) 28 /* page size == SNDRV_TRIDENT_PAGE_SIZE */ 29 #define ALIGN_PAGE_SIZE PAGE_SIZE /* minimum page size for allocation */ 31 /* fill TLB entrie(s) corresponding to page wit 51 set_tlb_bus(struct snd_trident *trident, int page, dma_addr_t addr) argument 58 set_silent_tlb(struct snd_trident *trident, int page) argument 80 set_tlb_bus(struct snd_trident *trident, int page, dma_addr_t addr) argument 90 set_silent_tlb(struct snd_trident *trident, int page) argument 111 int page, psize; local 163 int idx, page; local 207 int page; local 264 int page; local [all...] |
/linux-master/fs/squashfs/ |
H A D | file_direct.c | 21 /* Read separately compressed datablock directly into page cache */ 22 int squashfs_readpage_block(struct page *target_page, u64 block, int bsize, 34 struct page **page; local 43 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); 44 if (page == NULL) 49 page[i] = (n == target_page->index) ? target_page : 52 if (page[i] == NULL) 55 if (PageUptodate(page[i])) { 56 unlock_page(page[ [all...] |
/linux-master/arch/openrisc/mm/ |
H A D | cache.c | 19 static __always_inline void cache_loop(struct page *page, const unsigned int reg) argument 21 unsigned long paddr = page_to_pfn(page) << PAGE_SHIFT; 30 void local_dcache_page_flush(struct page *page) argument 32 cache_loop(page, SPR_DCBFR); 36 void local_icache_page_inv(struct page *page) argument 38 cache_loop(page, SPR_ICBIR);
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 51 struct mlx5_ib_user_db_page *page; local 56 list_for_each_entry(page, &context->db_page_list, list) 57 if ((current->mm == page->mm) && 58 (page->user_virt == (virt & PAGE_MASK))) 61 page = kmalloc(sizeof(*page), GFP_KERNEL); 62 if (!page) { 67 page->user_virt = (virt & PAGE_MASK); 68 page->refcnt = 0; 69 page [all...] |
/linux-master/net/ceph/ |
H A D | pagelist.c | 32 struct page *page = list_entry(pl->head.prev, struct page, lru); local 33 kunmap(page); 44 struct page *page = list_first_entry(&pl->head, struct page, local 46 list_del(&page->lru); 47 __free_page(page); 56 struct page *pag local 110 struct page *page = __page_cache_alloc(GFP_NOFS); local 124 struct page *page = list_first_entry(&pl->free_list, local 153 struct page *page; local [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | cacheflush.h | 13 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page 47 static inline void flush_dcache_page(struct page *page) argument 49 flush_dcache_folio(page_folio(page)); 54 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, 65 extern void __flush_anon_page(struct page *page, unsigned long); 68 struct page *page, unsigne 67 flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument [all...] |
H A D | tlbflush.h | 10 * - flush_tlb_page(vma, vmaddr) flushes one page 20 unsigned long page); 23 extern void local_flush_tlb_one(unsigned long asid, unsigned long page); 33 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 35 extern void flush_tlb_one(unsigned long asid, unsigned long page); 41 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) 42 #define flush_tlb_one(asid, page) local_flush_tlb_one(asid, page)
|
/linux-master/fs/sysv/ |
H A D | dir.c | 31 static void dir_commit_chunk(struct page *page, loff_t pos, unsigned len) argument 33 struct address_space *mapping = page->mapping; 36 block_write_end(NULL, mapping, pos, len, len, page, NULL); 41 unlock_page(page); 61 static void *dir_get_page(struct inode *dir, unsigned long n, struct page **p) 64 struct page *page = read_mapping_page(mapping, n, NULL); local 65 if (IS_ERR(page)) 66 return ERR_CAST(page); 90 struct page *page; local 146 struct page *page = NULL; local 189 struct page *page = NULL; local 238 sysv_delete_entry(struct sysv_dir_entry *de, struct page *page) argument 259 struct page *page = grab_cache_page(inode->i_mapping, 0); local 295 struct page *page = NULL; local 334 sysv_set_link(struct sysv_dir_entry *de, struct page *page, struct inode *inode) argument 373 struct page *page; local [all...] |