Searched refs:page (Results 226 - 250 of 3141) sorted by relevance

1234567891011>>

/linux-master/mm/kmsan/
H A Dinit.c28 * the page allocator becomes available.
97 struct page *shadow, *origin;
114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) argument
116 struct page *shadow, *origin;
119 held_back[order].shadow = page;
123 held_back[order].origin = page;
128 kmsan_setup_meta(page, shadow, origin, order);
137 struct page *items[MAX_BLOCKS];
147 static void smallstack_push(struct smallstack *stack, struct page *page
168 struct page *page, *shadow, *origin; local
185 struct page *page; local
[all...]
/linux-master/mm/
H A Dpercpu-vm.c8 * Chunks are mapped into vmalloc areas and populated page by page.
13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk,
25 * Returns pointer to array of pointers to struct page which can be indexed
32 static struct page **pcpu_get_pages(void)
34 static struct page **pages;
48 * @page_start: page index of the first page to be freed
49 * @page_end: page index of the last page t
62 struct page *page = pages[pcpu_page_idx(cpu, i)]; local
161 struct page *page; local
[all...]
H A Dshuffle.c36 static struct page * __meminit shuffle_valid_page(struct zone *zone,
39 struct page *page = pfn_to_online_page(pfn); local
46 /* ... is the page managed by the buddy? */
47 if (!page)
50 /* ... is the page assigned to the same zone? */
51 if (page_zone(page) != zone)
54 /* ...is the page free and currently on a free_area list? */
55 if (!PageBuddy(page))
59 * ...is the page o
[all...]
H A Ddmapool.c10 * the given device. It uses the dma_alloc_coherent page allocator to get
16 * allocated pages. Each page in the page_list is split into blocks of at
19 * keep a count of how many are currently allocated from each page.
122 struct dma_page *page; local
124 list_for_each_entry(page, &pool->page_list, page_list) {
125 if (dma < page->dma)
127 if ((dma - page->dma) < pool->allocation)
128 return page;
136 struct dma_page *page; local
138 page
159 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument
176 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument
300 pool_initialise_page(struct dma_pool *pool, struct dma_page *page) argument
336 struct dma_page *page; local
362 struct dma_page *page, *tmp; local
408 struct dma_page *page; local
[all...]
/linux-master/include/linux/
H A Dpagemap.h147 * mapping_shrinkable - test if page cache state allows inode reclaim
148 * @mapping: the page cache mapping
173 * inodes before there is highmem pressure from the page
186 * head pointer, which allows non-resident page cache entries
317 * reduce page locking.
425 struct address_space *page_mapping(struct page *);
433 * For folios which are in the page cache, return the mapping that this
434 * page belongs to. Folios in the swap cache return the mapping of the
439 * Do not call this for folios which aren't in the page cache or swap cache.
453 * For folios which are in the page cach
469 page_file_mapping(struct page *page) argument
544 attach_page_private(struct page *page, void *data) argument
549 detach_page_private(struct page *page) argument
908 page_to_pgoff(struct page *page) argument
926 page_offset(struct page *page) argument
931 page_file_offset(struct page *page) argument
1025 trylock_page(struct page *page) argument
1070 lock_page(struct page *page) argument
1141 wait_on_page_locked(struct page *page) argument
1386 struct page *page; local
1399 VM_BUG_ON_PAGE(PageTail(page), page); local
1511 page_mkwrite_check_truncate(struct page *page, struct inode *inode) argument
1549 i_blocks_per_page(struct inode *inode, struct page *page) argument
[all...]
/linux-master/kernel/power/
H A Dsnapshot.c83 * The calls to set_direct_map_*() should not fail because remapping a page
88 static inline void hibernate_map_page(struct page *page) argument
91 int ret = set_direct_map_default_noflush(page);
94 pr_warn_once("Failed to remap page\n");
96 debug_pagealloc_map_pages(page, 1);
100 static inline void hibernate_unmap_page(struct page *page) argument
103 unsigned long addr = (unsigned long)page_address(page);
104 int ret = set_direct_map_invalid_noflush(page);
229 struct page *page; local
257 struct page *page; local
1042 swsusp_set_page_free(struct page *page) argument
1048 swsusp_page_is_free(struct page *page) argument
1054 swsusp_unset_page_free(struct page *page) argument
1060 swsusp_set_page_forbidden(struct page *page) argument
1066 swsusp_page_is_forbidden(struct page *page) argument
1072 swsusp_unset_page_forbidden(struct page *page) argument
1190 clear_or_poison_free_page(struct page *page) argument
1253 struct page *page; local
1323 struct page *page; local
1387 struct page *page; local
1626 struct page *page = pfn_to_page(fr_pfn); local
1662 struct page *page; local
1762 struct page *page = pfn_to_page(pfn); local
2064 struct page *page; local
2103 struct page *page; local
2270 struct page *page; local
2478 struct page *page; local
2515 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument
2598 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument
2729 struct page *page; local
[all...]
/linux-master/drivers/s390/char/
H A Dsclp_con.c50 void *page; local
53 page = sclp_unmake_buffer(buffer);
58 list_add_tail((struct list_head *) page, &sclp_con_pages);
139 void *page; local
151 page = sclp_unmake_buffer(buffer);
152 list_add_tail((struct list_head *) page, &sclp_con_pages);
164 void *page; local
186 page = sclp_con_pages.next;
187 list_del((struct list_head *) page);
188 sclp_conbuf = sclp_make_buffer(page, SCLP_CON_COLUMN
272 void *page; local
[all...]
/linux-master/arch/m68k/include/asm/
H A Dcacheflush_mm.h222 /* Push the page at kernel virtual address and clear the icache */
258 #define flush_dcache_page(page) __flush_pages_to_ram(page_address(page), 1)
263 #define flush_icache_pages(vma, page, nr) \
264 __flush_pages_to_ram(page_address(page), nr)
266 extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page,
273 struct page *page, unsigned long vaddr,
276 flush_cache_page(vma, vaddr, page_to_pfn(page));
272 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
280 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
[all...]
/linux-master/arch/xtensa/include/asm/
H A Dpage.h2 * include/asm-xtensa/page.h
22 * PAGE_SHIFT determines the page size
43 * If the cache size for one way is greater than the page size, we have to
44 * deal with cache aliasing. The cache index is wider than the page size:
55 * When the page number is translated to the physical page address, the lowest
59 * The kernel does not provide a mechanism to ensure that the page color
62 * the page might also change.
99 typedef struct { unsigned long pte; } pte_t; /* page table entry */
102 typedef struct page *pgtable_
[all...]
/linux-master/tools/virtio/linux/
H A Dscatterlist.h21 * sg_assign_page - Assign a given page to an SG entry
23 * @page: The page
26 * Assign page to sg entry. Also see sg_set_page(), the most commonly used
30 static inline void sg_assign_page(struct scatterlist *sg, struct page *page) argument
38 BUG_ON((unsigned long) page & 0x03);
42 sg->page_link = page_link | (unsigned long) page;
46 * sg_set_page - Set sg entry to point at given page
48 * @page
59 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len, unsigned int offset) argument
[all...]
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_gem_object.c16 struct page *page; local
18 for_each_sgt_page(page, sgt_iter, pages) {
19 __free_page(page);
52 struct page *page; local
54 page = alloc_page(GFP | __GFP_HIGHMEM);
55 if (!page) {
60 sg_set_page(sg, page, PAGE_SIZE, 0);
/linux-master/drivers/hwmon/pmbus/
H A Dmax31785.c82 int page, int command)
87 rc = pmbus_read_byte_data(client, page, command);
94 int page, int command, u16 data)
99 rc = pmbus_write_byte_data(client, page, command, data);
106 int page, int phase, int command)
111 rc = pmbus_read_word_data(client, page, phase, command);
118 int page, int command, u16 data)
123 rc = pmbus_write_word_data(client, page, command, data);
128 static int max31785_read_byte_data(struct i2c_client *client, int page, int reg) argument
138 page
80 _max31785_read_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command) argument
92 _max31785_write_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument
104 _max31785_read_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int phase, int command) argument
116 _max31785_write_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument
145 max31785_write_byte(struct i2c_client *client, int page, u8 value) argument
153 max31785_read_long_data(struct i2c_client *client, int page, int reg, u32 *data) argument
191 max31785_get_pwm(struct i2c_client *client, int page) argument
206 max31785_get_pwm_mode(struct i2c_client *client, struct max31785_data *driver_data, int page) argument
233 max31785_read_word_data(struct i2c_client *client, int page, int phase, int reg) argument
297 max31785_update_fan(struct i2c_client *client, struct max31785_data *driver_data, int page, u8 config, u8 mask, u16 command) argument
324 max31785_pwm_enable(struct i2c_client *client, struct max31785_data *driver_data, int page, u16 word) argument
358 max31785_write_word_data(struct i2c_client *client, int page, int reg, u16 word) argument
[all...]
/linux-master/drivers/gpu/drm/gma500/
H A Dmmu.h13 * before taking the page table spinlock.
17 /* protects page tables, directory tables and pt tables.
39 struct page *p;
47 struct page *p;
48 struct page *dummy_pt;
49 struct page *dummy_page;
77 extern int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages,
/linux-master/drivers/usb/gadget/function/
H A Du_ether_configfs.h30 char *page) \
36 result = gether_get_dev_addr(opts->net, page, PAGE_SIZE); \
43 const char *page, size_t len)\
54 ret = gether_set_dev_addr(opts->net, page); \
65 char *page) \
71 result = gether_get_host_addr(opts->net, page, PAGE_SIZE); \
78 const char *page, size_t len)\
89 ret = gether_set_host_addr(opts->net, page); \
100 char *page) \
108 return sprintf(page, "
[all...]
/linux-master/tools/testing/nvdimm/
H A Dpmem-dax.c22 * Limit dax to a single page at a time given vmalloc()-backed
26 struct page *page; local
30 page = vmalloc_to_page(pmem->virt_addr + offset);
32 *pfn = page_to_pfn_t(page);
34 __func__, pmem, pgoff, page_to_pfn(page));
/linux-master/arch/arm/mm/
H A Dcopypage-xscale.c8 * processors. When we copy a user page page, we map it in such a way
9 * that accesses to this page will not touch the main data cache, but
11 * the main data cache on page faults.
84 void xscale_mc_copy_user_highpage(struct page *to, struct page *from,
108 xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument
110 void *ptr, *kaddr = kmap_atomic(page);
H A Dcopypage-v4mc.c8 * processors. When we copy a user page page, we map it in such a way
9 * that accesses to this page will not touch the main data cache, but
11 * the main data cache on page faults.
64 void v4_mc_copy_user_highpage(struct page *to, struct page *from,
87 void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument
89 void *ptr, *kaddr = kmap_atomic(page);
/linux-master/arch/csky/abiv2/inc/abi/
H A Dcacheflush.h29 static inline void flush_dcache_page(struct page *page) argument
31 flush_dcache_folio(page_folio(page));
47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
58 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/linux-master/arch/x86/include/asm/
H A Dpage.h19 struct page;
25 static inline void clear_user_page(void *page, unsigned long vaddr, argument
26 struct page *pg)
28 clear_page(page);
32 struct page *topage)
/linux-master/fs/f2fs/
H A Ddir.c166 struct page *dentry_page,
303 struct page **res_page)
308 struct page *dentry_page;
362 struct page **res_page)
401 * It returns the page where the entry was found (as a parameter - res_page),
406 const struct qstr *child, struct page **res_page)
427 struct f2fs_dir_entry *f2fs_parent_dir(struct inode *dir, struct page **p)
433 struct page **page)
438 de = f2fs_find_entry(dir, qstr, page);
432 f2fs_inode_by_name(struct inode *dir, const struct qstr *qstr, struct page **page) argument
447 f2fs_set_link(struct inode *dir, struct f2fs_dir_entry *de, struct page *page, struct inode *inode) argument
512 make_empty_dir(struct inode *inode, struct inode *parent, struct page *page) argument
539 struct page *page; local
689 struct page *page = NULL; local
802 struct page *page = NULL; local
836 struct page *page; local
881 f2fs_delete_entry(struct f2fs_dir_entry *dentry, struct page *page, struct inode *dir, struct inode *inode) argument
[all...]
H A Drecovery.c167 static int recover_dentry(struct inode *inode, struct page *ipage,
175 struct page *page; local
197 de = __f2fs_find_entry(dir, &fname, &page);
222 f2fs_delete_entry(de, page, dir, einode);
225 } else if (IS_ERR(page)) {
226 err = PTR_ERR(page);
236 f2fs_put_page(page, 0);
248 static int recover_quota_data(struct inode *inode, struct page *page) argument
287 recover_inode(struct inode *inode, struct page *page) argument
367 struct page *page = NULL; local
410 struct page *page = NULL; local
626 do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page) argument
782 struct page *page = NULL; local
[all...]
/linux-master/scripts/kconfig/lxdialog/
H A Dtextbox.c13 static const char *buf, *page; variable
18 * 'page' will be updated to point to the desired line in 'buf'.
27 if (*page == '\0') {
33 if (page == buf) {
37 page--;
39 if (page == buf) {
43 page--;
44 } while (*page != '\n');
45 page++;
51 * 'page' shoul
[all...]
/linux-master/arch/sh/mm/
H A Dnommu.c16 #include <asm/page.h>
55 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
60 void local_flush_tlb_one(unsigned long asid, unsigned long page) argument
82 void *kmap_coherent(struct page *page, unsigned long addr) argument
/linux-master/drivers/net/dsa/b53/
H A Db53_spi.c76 static inline int b53_spi_set_page(struct spi_device *spi, u8 page) argument
82 txbuf[2] = page;
87 static inline int b53_prepare_reg_access(struct spi_device *spi, u8 page) argument
94 return b53_spi_set_page(spi, page);
124 static int b53_spi_read(struct b53_device *dev, u8 page, u8 reg, u8 *data, argument
130 ret = b53_prepare_reg_access(spi, page);
141 static int b53_spi_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument
143 return b53_spi_read(dev, page, reg, val, 1);
146 static int b53_spi_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument
151 ret = b53_spi_read(dev, page, re
159 b53_spi_read32(struct b53_device *dev, u8 page, u8 reg, u32 *val) argument
172 b53_spi_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument
185 b53_spi_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument
198 b53_spi_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument
215 b53_spi_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument
232 b53_spi_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument
249 b53_spi_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument
266 b53_spi_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument
[all...]
/linux-master/drivers/char/agp/
H A Defficeon-agp.c28 #include <linux/page-flags.h>
38 * The page array is filled through the ATTPAGE register
40 * 31:20: physical page address
43 * mapped pages (the 2nd level page table pages
69 static inline unsigned long efficeon_mask_memory(struct page *page) argument
71 unsigned long addr = page_to_phys(page);
163 unsigned long page = efficeon_private.l1_table[index]; local
164 if (page) {
166 free_page(page);
210 unsigned long page; local
239 unsigned int *page, *last_page; local
300 unsigned int *page = (unsigned int *) efficeon_private.l1_table[index >> 10]; local
[all...]

Completed in 394 milliseconds

1234567891011>>