/linux-master/mm/ |
H A D | percpu-vm.c | 8 * Chunks are mapped into vmalloc areas and populated page by page. 13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, 25 * Returns pointer to array of pointers to struct page which can be indexed 32 static struct page **pcpu_get_pages(void) 34 static struct page **pages; 48 * @page_start: page index of the first page to be freed 49 * @page_end: page index of the last page t 62 struct page *page = pages[pcpu_page_idx(cpu, i)]; local 161 struct page *page; local [all...] |
H A D | gup.c | 33 static inline void sanity_check_pinned_pages(struct page **pages, 42 * stick around until the page is freed. 47 * THP we can assume that either the given page (PTE-mapped THP) or 48 * the head page (PMD-mapped THP) should be PageAnonExclusive(). If 52 struct page *page = *pages; local 53 struct folio *folio = page_folio(page); 55 if (is_zero_page(page) || 59 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page), page); 71 try_get_folio(struct page *page, int refs) argument 126 try_grab_folio(struct page *page, int refs, unsigned int flags) argument 225 try_grab_page(struct page *page, unsigned int flags) argument 272 unpin_user_page(struct page *page) argument 424 unpin_user_page_range_dirty_lock(struct page *page, unsigned long npages, bool make_dirty) argument 542 can_follow_write_pte(pte_t pte, struct page *page, struct vm_area_struct *vma, unsigned int flags) argument 584 struct page *page; local 698 struct page *page; local 750 struct page *page; local 841 struct page *page; local 859 get_gate_page(struct mm_struct *mm, unsigned long address, unsigned int gup_flags, struct vm_area_struct **vma, struct page **page) argument 1203 struct page *page; local 2039 struct page *page; local 2549 struct page *page = pages[--(*nr)]; local 2592 struct page *page; local 2702 struct page *page = pfn_to_page(pfn); local 2782 record_subpages(struct page *page, unsigned long addr, unsigned long end, struct page **pages) argument 2806 struct page *page; local 2880 struct page *page; local 2924 struct page *page; local 2970 struct page *page; local [all...] |
H A D | shuffle.c | 36 static struct page * __meminit shuffle_valid_page(struct zone *zone, 39 struct page *page = pfn_to_online_page(pfn); local 46 /* ... is the page managed by the buddy? */ 47 if (!page) 50 /* ... is the page assigned to the same zone? */ 51 if (page_zone(page) != zone) 54 /* ...is the page free and currently on a free_area list? */ 55 if (!PageBuddy(page)) 59 * ...is the page o [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_db.c | 14 struct hns_roce_user_db_page *page; local 20 list_for_each_entry(page, &context->page_list, list) 21 if (page->user_virt == page_addr) 24 page = kmalloc(sizeof(*page), GFP_KERNEL); 25 if (!page) { 30 refcount_set(&page->refcount, 1); 31 page->user_virt = page_addr; 32 page->umem = ib_umem_get(context->ibucontext.device, page_addr, 34 if (IS_ERR(page [all...] |
/linux-master/fs/f2fs/ |
H A D | inline.c | 64 void f2fs_do_read_inline_data(struct page *page, struct page *ipage) argument 66 struct inode *inode = page->mapping->host; 68 if (PageUptodate(page)) 71 f2fs_bug_on(F2FS_P_SB(page), page->index); 73 zero_user_segment(page, MAX_INLINE_DATA(inode), PAGE_SIZE); 76 memcpy_to_page(page, 0, inline_data_addr(inode, ipage), 78 if (!PageUptodate(page)) 100 f2fs_read_inline_data(struct inode *inode, struct page *page) argument 127 f2fs_convert_inline_page(struct dnode_of_data *dn, struct page *page) argument 203 struct page *ipage, *page; local 243 f2fs_write_inline_data(struct inode *inode, struct page *page) argument 262 page, 0, MAX_INLINE_DATA(inode)); local 393 struct page *page; local 624 struct page *page = NULL; local 680 f2fs_delete_inline_entry(struct f2fs_dir_entry *dentry, struct page *page, struct inode *dir, struct inode *inode) argument [all...] |
H A D | data.c | 51 bool f2fs_is_cp_guaranteed(struct page *page) argument 53 struct address_space *mapping = page->mapping; 69 page_private_gcing(page)) 74 static enum count_type __read_io_type(struct page *page) argument 76 struct address_space *mapping = page_file_mapping(page); 133 * things for each compressed page here: call f2fs_end_read_compressed_page() 136 * release the bio's reference to the decompress_io_ctx of the page's cluster. 145 struct page *pag local 194 struct page *page = bv->bv_page; local 248 struct page *page = bv->bv_page; local 338 struct page *page = bvec->bv_page; local 544 __has_merged_page(struct bio *bio, struct inode *inode, struct page *page, nid_t ino) argument 637 __submit_merged_write_cond(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page, nid_t ino, enum page_type type, bool force) argument 667 f2fs_submit_merged_write_cond(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page, nid_t ino, enum page_type type) argument 688 struct page *page = fio->encrypted_page ? local 752 add_bio_entry(struct f2fs_sb_info *sbi, struct bio *bio, struct page *page, enum temp_type temp) argument 776 add_ipu_page(struct f2fs_io_info *fio, struct bio **bio, struct page *page) argument 824 f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi, struct bio **bio, struct page *page) argument 885 struct page *page = fio->encrypted_page ? local 1083 f2fs_submit_page_read(struct inode *inode, struct page *page, block_t blkaddr, blk_opf_t op_flags, bool for_write) argument 1206 struct page *page; local 1281 struct page *page; local 1312 struct page *page; local 1340 struct page *page; local 1801 struct page *page; local 2045 f2fs_read_single_page(struct inode *inode, struct page *page, unsigned nr_pages, struct f2fs_map_blocks *map, struct bio **bio_ret, sector_t *last_block_in_bio, bool is_readahead) argument 2177 struct page *page = cc->rpages[i]; local 2250 struct page *page = dic->cpages[i]; local 2326 f2fs_mpage_readpages(struct inode *inode, struct readahead_control *rac, struct page *page) argument 2439 struct page *page = &folio->page; local 2477 struct page *mpage, *page; local 2616 struct page *page = fio->page; local 2732 f2fs_write_single_data_page(struct page *page, int *submitted, struct bio **bio, sector_t *last_block, struct writeback_control *wbc, enum iostat_type io_type, int compr_blocks, bool allow_balance) argument 2905 f2fs_write_data_page(struct page *page, struct writeback_control *wbc) argument 3039 struct page *page = pages[i]; local 3340 prepare_write_begin(struct f2fs_sb_info *sbi, struct page *page, loff_t pos, unsigned len, block_t *blk_addr, bool *node_changed) argument 3480 prepare_atomic_write_begin(struct f2fs_sb_info *sbi, struct page *page, loff_t pos, unsigned int len, block_t *blk_addr, bool *node_changed, bool *use_cow) argument 3528 struct page *page = NULL; local 3654 f2fs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument 3858 struct page *page; local 4083 f2fs_clear_page_cache_dirty_tag(struct page *page) argument [all...] |
/linux-master/kernel/dma/ |
H A D | pool.c | 83 struct page *page = NULL; local 93 page = dma_alloc_from_contiguous(NULL, 1 << order, 95 if (!page) 96 page = alloc_pages(gfp, order); 97 } while (!page && order-- > 0); 98 if (!page) 101 arch_dma_prep_coherent(page, pool_size); 104 addr = dma_common_contiguous_remap(page, pool_size, 110 addr = page_to_virt(page); 270 struct page *page; local [all...] |
/linux-master/mm/kmsan/ |
H A D | init.c | 28 * the page allocator becomes available. 97 struct page *shadow, *origin; 114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) argument 116 struct page *shadow, *origin; 119 held_back[order].shadow = page; 123 held_back[order].origin = page; 128 kmsan_setup_meta(page, shadow, origin, order); 137 struct page *items[MAX_BLOCKS]; 147 static void smallstack_push(struct smallstack *stack, struct page *page 168 struct page *page, *shadow, *origin; local 185 struct page *page; local [all...] |
/linux-master/kernel/power/ |
H A D | snapshot.c | 83 * The calls to set_direct_map_*() should not fail because remapping a page 88 static inline void hibernate_map_page(struct page *page) argument 91 int ret = set_direct_map_default_noflush(page); 94 pr_warn_once("Failed to remap page\n"); 96 debug_pagealloc_map_pages(page, 1); 100 static inline void hibernate_unmap_page(struct page *page) argument 103 unsigned long addr = (unsigned long)page_address(page); 104 int ret = set_direct_map_invalid_noflush(page); 229 struct page *page; local 257 struct page *page; local 1042 swsusp_set_page_free(struct page *page) argument 1048 swsusp_page_is_free(struct page *page) argument 1054 swsusp_unset_page_free(struct page *page) argument 1060 swsusp_set_page_forbidden(struct page *page) argument 1066 swsusp_page_is_forbidden(struct page *page) argument 1072 swsusp_unset_page_forbidden(struct page *page) argument 1190 clear_or_poison_free_page(struct page *page) argument 1253 struct page *page; local 1323 struct page *page; local 1387 struct page *page; local 1626 struct page *page = pfn_to_page(fr_pfn); local 1662 struct page *page; local 1762 struct page *page = pfn_to_page(pfn); local 2064 struct page *page; local 2103 struct page *page; local 2270 struct page *page; local 2478 struct page *page; local 2515 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument 2598 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument 2729 struct page *page; local [all...] |
/linux-master/fs/proc/ |
H A D | page.c | 17 #include <linux/kernel-page-flags.h> 40 /* /proc/kpagecount - an array exposing page counts 43 * physical page count. 50 struct page *ppage; 99 /* /proc/kpageflags - an array exposing page flags 102 * physical page flags. 110 u64 stable_page_flags(struct page *page) argument 117 * it differentiates a memory hole from a page with no flags 119 if (!page) [all...] |
/linux-master/drivers/s390/char/ |
H A D | sclp_con.c | 50 void *page; local 53 page = sclp_unmake_buffer(buffer); 58 list_add_tail((struct list_head *) page, &sclp_con_pages); 139 void *page; local 151 page = sclp_unmake_buffer(buffer); 152 list_add_tail((struct list_head *) page, &sclp_con_pages); 164 void *page; local 186 page = sclp_con_pages.next; 187 list_del((struct list_head *) page); 188 sclp_conbuf = sclp_make_buffer(page, SCLP_CON_COLUMN 272 void *page; local [all...] |
/linux-master/arch/m68k/include/asm/ |
H A D | cacheflush_mm.h | 222 /* Push the page at kernel virtual address and clear the icache */ 258 #define flush_dcache_page(page) __flush_pages_to_ram(page_address(page), 1) 263 #define flush_icache_pages(vma, page, nr) \ 264 __flush_pages_to_ram(page_address(page), nr) 266 extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page, 273 struct page *page, unsigned long vaddr, 276 flush_cache_page(vma, vaddr, page_to_pfn(page)); 272 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument 280 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument [all...] |
/linux-master/arch/xtensa/include/asm/ |
H A D | page.h | 2 * include/asm-xtensa/page.h 22 * PAGE_SHIFT determines the page size 43 * If the cache size for one way is greater than the page size, we have to 44 * deal with cache aliasing. The cache index is wider than the page size: 55 * When the page number is translated to the physical page address, the lowest 59 * The kernel does not provide a mechanism to ensure that the page color 62 * the page might also change. 99 typedef struct { unsigned long pte; } pte_t; /* page table entry */ 102 typedef struct page *pgtable_ [all...] |
H A D | cacheflush.h | 14 #include <asm/page.h> 89 * Pages can get remapped. Because this might change the 'color' of that page, 127 static inline void flush_dcache_page(struct page *page) argument 129 flush_dcache_folio(page_folio(page)); 148 #define flush_dcache_page(page) do { } while (0) 170 extern void copy_to_user_page(struct vm_area_struct*, struct page*, 172 extern void copy_from_user_page(struct vm_area_struct*, struct page*, 177 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 184 #define copy_from_user_page(vma, page, vadd [all...] |
/linux-master/tools/virtio/linux/ |
H A D | scatterlist.h | 21 * sg_assign_page - Assign a given page to an SG entry 23 * @page: The page 26 * Assign page to sg entry. Also see sg_set_page(), the most commonly used 30 static inline void sg_assign_page(struct scatterlist *sg, struct page *page) argument 38 BUG_ON((unsigned long) page & 0x03); 42 sg->page_link = page_link | (unsigned long) page; 46 * sg_set_page - Set sg entry to point at given page 48 * @page 59 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len, unsigned int offset) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 16 struct page *page; local 18 for_each_sgt_page(page, sgt_iter, pages) { 19 __free_page(page); 52 struct page *page; local 54 page = alloc_page(GFP | __GFP_HIGHMEM); 55 if (!page) { 60 sg_set_page(sg, page, PAGE_SIZE, 0);
|
/linux-master/drivers/hwmon/pmbus/ |
H A D | max31785.c | 82 int page, int command) 87 rc = pmbus_read_byte_data(client, page, command); 94 int page, int command, u16 data) 99 rc = pmbus_write_byte_data(client, page, command, data); 106 int page, int phase, int command) 111 rc = pmbus_read_word_data(client, page, phase, command); 118 int page, int command, u16 data) 123 rc = pmbus_write_word_data(client, page, command, data); 128 static int max31785_read_byte_data(struct i2c_client *client, int page, int reg) argument 138 page 80 _max31785_read_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command) argument 92 _max31785_write_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument 104 _max31785_read_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int phase, int command) argument 116 _max31785_write_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument 145 max31785_write_byte(struct i2c_client *client, int page, u8 value) argument 153 max31785_read_long_data(struct i2c_client *client, int page, int reg, u32 *data) argument 191 max31785_get_pwm(struct i2c_client *client, int page) argument 206 max31785_get_pwm_mode(struct i2c_client *client, struct max31785_data *driver_data, int page) argument 233 max31785_read_word_data(struct i2c_client *client, int page, int phase, int reg) argument 297 max31785_update_fan(struct i2c_client *client, struct max31785_data *driver_data, int page, u8 config, u8 mask, u16 command) argument 324 max31785_pwm_enable(struct i2c_client *client, struct max31785_data *driver_data, int page, u16 word) argument 358 max31785_write_word_data(struct i2c_client *client, int page, int reg, u16 word) argument [all...] |
/linux-master/drivers/gpu/drm/gma500/ |
H A D | mmu.h | 13 * before taking the page table spinlock. 17 /* protects page tables, directory tables and pt tables. 39 struct page *p; 47 struct page *p; 48 struct page *dummy_pt; 49 struct page *dummy_page; 77 extern int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages,
|
/linux-master/drivers/usb/gadget/function/ |
H A D | u_ether_configfs.h | 30 char *page) \ 36 result = gether_get_dev_addr(opts->net, page, PAGE_SIZE); \ 43 const char *page, size_t len)\ 54 ret = gether_set_dev_addr(opts->net, page); \ 65 char *page) \ 71 result = gether_get_host_addr(opts->net, page, PAGE_SIZE); \ 78 const char *page, size_t len)\ 89 ret = gether_set_host_addr(opts->net, page); \ 100 char *page) \ 108 return sprintf(page, " [all...] |
/linux-master/arch/sh/mm/ |
H A D | nommu.c | 14 #include <asm/page.h> 53 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 58 void local_flush_tlb_one(unsigned long asid, unsigned long page) argument 80 void *kmap_coherent(struct page *page, unsigned long addr) argument
|
/linux-master/tools/testing/nvdimm/ |
H A D | pmem-dax.c | 22 * Limit dax to a single page at a time given vmalloc()-backed 26 struct page *page; local 30 page = vmalloc_to_page(pmem->virt_addr + offset); 32 *pfn = page_to_pfn_t(page); 34 __func__, pmem, pgoff, page_to_pfn(page));
|
/linux-master/arch/arm/mm/ |
H A D | copypage-xscale.c | 8 * processors. When we copy a user page page, we map it in such a way 9 * that accesses to this page will not touch the main data cache, but 11 * the main data cache on page faults. 84 void xscale_mc_copy_user_highpage(struct page *to, struct page *from, 108 xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 110 void *ptr, *kaddr = kmap_atomic(page);
|
H A D | copypage-v4mc.c | 8 * processors. When we copy a user page page, we map it in such a way 9 * that accesses to this page will not touch the main data cache, but 11 * the main data cache on page faults. 64 void v4_mc_copy_user_highpage(struct page *to, struct page *from, 87 void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 89 void *ptr, *kaddr = kmap_atomic(page);
|
/linux-master/arch/csky/abiv2/inc/abi/ |
H A D | cacheflush.h | 29 static inline void flush_dcache_page(struct page *page) argument 31 flush_dcache_folio(page_folio(page)); 47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 58 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/arch/x86/include/asm/ |
H A D | page.h | 19 struct page; 25 static inline void clear_user_page(void *page, unsigned long vaddr, argument 26 struct page *pg) 28 clear_page(page); 32 struct page *topage)
|