Searched refs:page (Results 226 - 250 of 3144) sorted by relevance

1234567891011>>

/linux-master/mm/
H A Dpercpu-vm.c8 * Chunks are mapped into vmalloc areas and populated page by page.
13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk,
25 * Returns pointer to array of pointers to struct page which can be indexed
32 static struct page **pcpu_get_pages(void)
34 static struct page **pages;
48 * @page_start: page index of the first page to be freed
49 * @page_end: page index of the last page t
62 struct page *page = pages[pcpu_page_idx(cpu, i)]; local
161 struct page *page; local
[all...]
H A Dgup.c33 static inline void sanity_check_pinned_pages(struct page **pages,
42 * stick around until the page is freed.
47 * THP we can assume that either the given page (PTE-mapped THP) or
48 * the head page (PMD-mapped THP) should be PageAnonExclusive(). If
52 struct page *page = *pages; local
53 struct folio *folio = page_folio(page);
55 if (is_zero_page(page) ||
59 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page), page);
71 try_get_folio(struct page *page, int refs) argument
126 try_grab_folio(struct page *page, int refs, unsigned int flags) argument
225 try_grab_page(struct page *page, unsigned int flags) argument
272 unpin_user_page(struct page *page) argument
424 unpin_user_page_range_dirty_lock(struct page *page, unsigned long npages, bool make_dirty) argument
542 can_follow_write_pte(pte_t pte, struct page *page, struct vm_area_struct *vma, unsigned int flags) argument
584 struct page *page; local
698 struct page *page; local
750 struct page *page; local
841 struct page *page; local
859 get_gate_page(struct mm_struct *mm, unsigned long address, unsigned int gup_flags, struct vm_area_struct **vma, struct page **page) argument
1203 struct page *page; local
2039 struct page *page; local
2549 struct page *page = pages[--(*nr)]; local
2592 struct page *page; local
2702 struct page *page = pfn_to_page(pfn); local
2782 record_subpages(struct page *page, unsigned long addr, unsigned long end, struct page **pages) argument
2806 struct page *page; local
2880 struct page *page; local
2924 struct page *page; local
2970 struct page *page; local
[all...]
H A Dshuffle.c36 static struct page * __meminit shuffle_valid_page(struct zone *zone,
39 struct page *page = pfn_to_online_page(pfn); local
46 /* ... is the page managed by the buddy? */
47 if (!page)
50 /* ... is the page assigned to the same zone? */
51 if (page_zone(page) != zone)
54 /* ...is the page free and currently on a free_area list? */
55 if (!PageBuddy(page))
59 * ...is the page o
[all...]
/linux-master/drivers/infiniband/hw/hns/
H A Dhns_roce_db.c14 struct hns_roce_user_db_page *page; local
20 list_for_each_entry(page, &context->page_list, list)
21 if (page->user_virt == page_addr)
24 page = kmalloc(sizeof(*page), GFP_KERNEL);
25 if (!page) {
30 refcount_set(&page->refcount, 1);
31 page->user_virt = page_addr;
32 page->umem = ib_umem_get(context->ibucontext.device, page_addr,
34 if (IS_ERR(page
[all...]
/linux-master/fs/f2fs/
H A Dinline.c64 void f2fs_do_read_inline_data(struct page *page, struct page *ipage) argument
66 struct inode *inode = page->mapping->host;
68 if (PageUptodate(page))
71 f2fs_bug_on(F2FS_P_SB(page), page->index);
73 zero_user_segment(page, MAX_INLINE_DATA(inode), PAGE_SIZE);
76 memcpy_to_page(page, 0, inline_data_addr(inode, ipage),
78 if (!PageUptodate(page))
100 f2fs_read_inline_data(struct inode *inode, struct page *page) argument
127 f2fs_convert_inline_page(struct dnode_of_data *dn, struct page *page) argument
203 struct page *ipage, *page; local
243 f2fs_write_inline_data(struct inode *inode, struct page *page) argument
262 page, 0, MAX_INLINE_DATA(inode)); local
393 struct page *page; local
624 struct page *page = NULL; local
680 f2fs_delete_inline_entry(struct f2fs_dir_entry *dentry, struct page *page, struct inode *dir, struct inode *inode) argument
[all...]
H A Ddata.c51 bool f2fs_is_cp_guaranteed(struct page *page) argument
53 struct address_space *mapping = page->mapping;
69 page_private_gcing(page))
74 static enum count_type __read_io_type(struct page *page) argument
76 struct address_space *mapping = page_file_mapping(page);
133 * things for each compressed page here: call f2fs_end_read_compressed_page()
136 * release the bio's reference to the decompress_io_ctx of the page's cluster.
145 struct page *pag local
194 struct page *page = bv->bv_page; local
248 struct page *page = bv->bv_page; local
338 struct page *page = bvec->bv_page; local
544 __has_merged_page(struct bio *bio, struct inode *inode, struct page *page, nid_t ino) argument
637 __submit_merged_write_cond(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page, nid_t ino, enum page_type type, bool force) argument
667 f2fs_submit_merged_write_cond(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page, nid_t ino, enum page_type type) argument
688 struct page *page = fio->encrypted_page ? local
752 add_bio_entry(struct f2fs_sb_info *sbi, struct bio *bio, struct page *page, enum temp_type temp) argument
776 add_ipu_page(struct f2fs_io_info *fio, struct bio **bio, struct page *page) argument
824 f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi, struct bio **bio, struct page *page) argument
885 struct page *page = fio->encrypted_page ? local
1083 f2fs_submit_page_read(struct inode *inode, struct page *page, block_t blkaddr, blk_opf_t op_flags, bool for_write) argument
1206 struct page *page; local
1281 struct page *page; local
1312 struct page *page; local
1340 struct page *page; local
1801 struct page *page; local
2045 f2fs_read_single_page(struct inode *inode, struct page *page, unsigned nr_pages, struct f2fs_map_blocks *map, struct bio **bio_ret, sector_t *last_block_in_bio, bool is_readahead) argument
2177 struct page *page = cc->rpages[i]; local
2250 struct page *page = dic->cpages[i]; local
2326 f2fs_mpage_readpages(struct inode *inode, struct readahead_control *rac, struct page *page) argument
2439 struct page *page = &folio->page; local
2477 struct page *mpage, *page; local
2616 struct page *page = fio->page; local
2732 f2fs_write_single_data_page(struct page *page, int *submitted, struct bio **bio, sector_t *last_block, struct writeback_control *wbc, enum iostat_type io_type, int compr_blocks, bool allow_balance) argument
2905 f2fs_write_data_page(struct page *page, struct writeback_control *wbc) argument
3039 struct page *page = pages[i]; local
3340 prepare_write_begin(struct f2fs_sb_info *sbi, struct page *page, loff_t pos, unsigned len, block_t *blk_addr, bool *node_changed) argument
3480 prepare_atomic_write_begin(struct f2fs_sb_info *sbi, struct page *page, loff_t pos, unsigned int len, block_t *blk_addr, bool *node_changed, bool *use_cow) argument
3528 struct page *page = NULL; local
3654 f2fs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument
3858 struct page *page; local
4083 f2fs_clear_page_cache_dirty_tag(struct page *page) argument
[all...]
/linux-master/kernel/dma/
H A Dpool.c83 struct page *page = NULL; local
93 page = dma_alloc_from_contiguous(NULL, 1 << order,
95 if (!page)
96 page = alloc_pages(gfp, order);
97 } while (!page && order-- > 0);
98 if (!page)
101 arch_dma_prep_coherent(page, pool_size);
104 addr = dma_common_contiguous_remap(page, pool_size,
110 addr = page_to_virt(page);
270 struct page *page; local
[all...]
/linux-master/mm/kmsan/
H A Dinit.c28 * the page allocator becomes available.
97 struct page *shadow, *origin;
114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) argument
116 struct page *shadow, *origin;
119 held_back[order].shadow = page;
123 held_back[order].origin = page;
128 kmsan_setup_meta(page, shadow, origin, order);
137 struct page *items[MAX_BLOCKS];
147 static void smallstack_push(struct smallstack *stack, struct page *page
168 struct page *page, *shadow, *origin; local
185 struct page *page; local
[all...]
/linux-master/kernel/power/
H A Dsnapshot.c83 * The calls to set_direct_map_*() should not fail because remapping a page
88 static inline void hibernate_map_page(struct page *page) argument
91 int ret = set_direct_map_default_noflush(page);
94 pr_warn_once("Failed to remap page\n");
96 debug_pagealloc_map_pages(page, 1);
100 static inline void hibernate_unmap_page(struct page *page) argument
103 unsigned long addr = (unsigned long)page_address(page);
104 int ret = set_direct_map_invalid_noflush(page);
229 struct page *page; local
257 struct page *page; local
1042 swsusp_set_page_free(struct page *page) argument
1048 swsusp_page_is_free(struct page *page) argument
1054 swsusp_unset_page_free(struct page *page) argument
1060 swsusp_set_page_forbidden(struct page *page) argument
1066 swsusp_page_is_forbidden(struct page *page) argument
1072 swsusp_unset_page_forbidden(struct page *page) argument
1190 clear_or_poison_free_page(struct page *page) argument
1253 struct page *page; local
1323 struct page *page; local
1387 struct page *page; local
1626 struct page *page = pfn_to_page(fr_pfn); local
1662 struct page *page; local
1762 struct page *page = pfn_to_page(pfn); local
2064 struct page *page; local
2103 struct page *page; local
2270 struct page *page; local
2478 struct page *page; local
2515 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument
2598 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument
2729 struct page *page; local
[all...]
/linux-master/fs/proc/
H A Dpage.c17 #include <linux/kernel-page-flags.h>
40 /* /proc/kpagecount - an array exposing page counts
43 * physical page count.
50 struct page *ppage;
99 /* /proc/kpageflags - an array exposing page flags
102 * physical page flags.
110 u64 stable_page_flags(struct page *page) argument
117 * it differentiates a memory hole from a page with no flags
119 if (!page)
[all...]
/linux-master/drivers/s390/char/
H A Dsclp_con.c50 void *page; local
53 page = sclp_unmake_buffer(buffer);
58 list_add_tail((struct list_head *) page, &sclp_con_pages);
139 void *page; local
151 page = sclp_unmake_buffer(buffer);
152 list_add_tail((struct list_head *) page, &sclp_con_pages);
164 void *page; local
186 page = sclp_con_pages.next;
187 list_del((struct list_head *) page);
188 sclp_conbuf = sclp_make_buffer(page, SCLP_CON_COLUMN
272 void *page; local
[all...]
/linux-master/arch/m68k/include/asm/
H A Dcacheflush_mm.h222 /* Push the page at kernel virtual address and clear the icache */
258 #define flush_dcache_page(page) __flush_pages_to_ram(page_address(page), 1)
263 #define flush_icache_pages(vma, page, nr) \
264 __flush_pages_to_ram(page_address(page), nr)
266 extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page,
273 struct page *page, unsigned long vaddr,
276 flush_cache_page(vma, vaddr, page_to_pfn(page));
272 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
280 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
[all...]
/linux-master/arch/xtensa/include/asm/
H A Dpage.h2 * include/asm-xtensa/page.h
22 * PAGE_SHIFT determines the page size
43 * If the cache size for one way is greater than the page size, we have to
44 * deal with cache aliasing. The cache index is wider than the page size:
55 * When the page number is translated to the physical page address, the lowest
59 * The kernel does not provide a mechanism to ensure that the page color
62 * the page might also change.
99 typedef struct { unsigned long pte; } pte_t; /* page table entry */
102 typedef struct page *pgtable_
[all...]
H A Dcacheflush.h14 #include <asm/page.h>
89 * Pages can get remapped. Because this might change the 'color' of that page,
127 static inline void flush_dcache_page(struct page *page) argument
129 flush_dcache_folio(page_folio(page));
148 #define flush_dcache_page(page) do { } while (0)
170 extern void copy_to_user_page(struct vm_area_struct*, struct page*,
172 extern void copy_from_user_page(struct vm_area_struct*, struct page*,
177 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
184 #define copy_from_user_page(vma, page, vadd
[all...]
/linux-master/tools/virtio/linux/
H A Dscatterlist.h21 * sg_assign_page - Assign a given page to an SG entry
23 * @page: The page
26 * Assign page to sg entry. Also see sg_set_page(), the most commonly used
30 static inline void sg_assign_page(struct scatterlist *sg, struct page *page) argument
38 BUG_ON((unsigned long) page & 0x03);
42 sg->page_link = page_link | (unsigned long) page;
46 * sg_set_page - Set sg entry to point at given page
48 * @page
59 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len, unsigned int offset) argument
[all...]
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_gem_object.c16 struct page *page; local
18 for_each_sgt_page(page, sgt_iter, pages) {
19 __free_page(page);
52 struct page *page; local
54 page = alloc_page(GFP | __GFP_HIGHMEM);
55 if (!page) {
60 sg_set_page(sg, page, PAGE_SIZE, 0);
/linux-master/drivers/hwmon/pmbus/
H A Dmax31785.c82 int page, int command)
87 rc = pmbus_read_byte_data(client, page, command);
94 int page, int command, u16 data)
99 rc = pmbus_write_byte_data(client, page, command, data);
106 int page, int phase, int command)
111 rc = pmbus_read_word_data(client, page, phase, command);
118 int page, int command, u16 data)
123 rc = pmbus_write_word_data(client, page, command, data);
128 static int max31785_read_byte_data(struct i2c_client *client, int page, int reg) argument
138 page
80 _max31785_read_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command) argument
92 _max31785_write_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument
104 _max31785_read_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int phase, int command) argument
116 _max31785_write_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument
145 max31785_write_byte(struct i2c_client *client, int page, u8 value) argument
153 max31785_read_long_data(struct i2c_client *client, int page, int reg, u32 *data) argument
191 max31785_get_pwm(struct i2c_client *client, int page) argument
206 max31785_get_pwm_mode(struct i2c_client *client, struct max31785_data *driver_data, int page) argument
233 max31785_read_word_data(struct i2c_client *client, int page, int phase, int reg) argument
297 max31785_update_fan(struct i2c_client *client, struct max31785_data *driver_data, int page, u8 config, u8 mask, u16 command) argument
324 max31785_pwm_enable(struct i2c_client *client, struct max31785_data *driver_data, int page, u16 word) argument
358 max31785_write_word_data(struct i2c_client *client, int page, int reg, u16 word) argument
[all...]
/linux-master/drivers/gpu/drm/gma500/
H A Dmmu.h13 * before taking the page table spinlock.
17 /* protects page tables, directory tables and pt tables.
39 struct page *p;
47 struct page *p;
48 struct page *dummy_pt;
49 struct page *dummy_page;
77 extern int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages,
/linux-master/drivers/usb/gadget/function/
H A Du_ether_configfs.h30 char *page) \
36 result = gether_get_dev_addr(opts->net, page, PAGE_SIZE); \
43 const char *page, size_t len)\
54 ret = gether_set_dev_addr(opts->net, page); \
65 char *page) \
71 result = gether_get_host_addr(opts->net, page, PAGE_SIZE); \
78 const char *page, size_t len)\
89 ret = gether_set_host_addr(opts->net, page); \
100 char *page) \
108 return sprintf(page, "
[all...]
/linux-master/arch/sh/mm/
H A Dnommu.c14 #include <asm/page.h>
53 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
58 void local_flush_tlb_one(unsigned long asid, unsigned long page) argument
80 void *kmap_coherent(struct page *page, unsigned long addr) argument
/linux-master/tools/testing/nvdimm/
H A Dpmem-dax.c22 * Limit dax to a single page at a time given vmalloc()-backed
26 struct page *page; local
30 page = vmalloc_to_page(pmem->virt_addr + offset);
32 *pfn = page_to_pfn_t(page);
34 __func__, pmem, pgoff, page_to_pfn(page));
/linux-master/arch/arm/mm/
H A Dcopypage-xscale.c8 * processors. When we copy a user page page, we map it in such a way
9 * that accesses to this page will not touch the main data cache, but
11 * the main data cache on page faults.
84 void xscale_mc_copy_user_highpage(struct page *to, struct page *from,
108 xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument
110 void *ptr, *kaddr = kmap_atomic(page);
H A Dcopypage-v4mc.c8 * processors. When we copy a user page page, we map it in such a way
9 * that accesses to this page will not touch the main data cache, but
11 * the main data cache on page faults.
64 void v4_mc_copy_user_highpage(struct page *to, struct page *from,
87 void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument
89 void *ptr, *kaddr = kmap_atomic(page);
/linux-master/arch/csky/abiv2/inc/abi/
H A Dcacheflush.h29 static inline void flush_dcache_page(struct page *page) argument
31 flush_dcache_folio(page_folio(page));
47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
58 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/linux-master/arch/x86/include/asm/
H A Dpage.h19 struct page;
25 static inline void clear_user_page(void *page, unsigned long vaddr, argument
26 struct page *pg)
28 clear_page(page);
32 struct page *topage)

Completed in 258 milliseconds

1234567891011>>