/linux-master/mm/kmsan/ |
H A D | init.c | 28 * the page allocator becomes available. 97 struct page *shadow, *origin; 114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) argument 116 struct page *shadow, *origin; 119 held_back[order].shadow = page; 123 held_back[order].origin = page; 128 kmsan_setup_meta(page, shadow, origin, order); 137 struct page *items[MAX_BLOCKS]; 147 static void smallstack_push(struct smallstack *stack, struct page *page 168 struct page *page, *shadow, *origin; local 185 struct page *page; local [all...] |
/linux-master/mm/ |
H A D | percpu-vm.c | 8 * Chunks are mapped into vmalloc areas and populated page by page. 13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, 25 * Returns pointer to array of pointers to struct page which can be indexed 32 static struct page **pcpu_get_pages(void) 34 static struct page **pages; 48 * @page_start: page index of the first page to be freed 49 * @page_end: page index of the last page t 62 struct page *page = pages[pcpu_page_idx(cpu, i)]; local 161 struct page *page; local [all...] |
H A D | shuffle.c | 36 static struct page * __meminit shuffle_valid_page(struct zone *zone, 39 struct page *page = pfn_to_online_page(pfn); local 46 /* ... is the page managed by the buddy? */ 47 if (!page) 50 /* ... is the page assigned to the same zone? */ 51 if (page_zone(page) != zone) 54 /* ...is the page free and currently on a free_area list? */ 55 if (!PageBuddy(page)) 59 * ...is the page o [all...] |
H A D | dmapool.c | 10 * the given device. It uses the dma_alloc_coherent page allocator to get 16 * allocated pages. Each page in the page_list is split into blocks of at 19 * keep a count of how many are currently allocated from each page. 122 struct dma_page *page; local 124 list_for_each_entry(page, &pool->page_list, page_list) { 125 if (dma < page->dma) 127 if ((dma - page->dma) < pool->allocation) 128 return page; 136 struct dma_page *page; local 138 page 159 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument 176 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument 300 pool_initialise_page(struct dma_pool *pool, struct dma_page *page) argument 336 struct dma_page *page; local 362 struct dma_page *page, *tmp; local 408 struct dma_page *page; local [all...] |
/linux-master/include/linux/ |
H A D | pagemap.h | 147 * mapping_shrinkable - test if page cache state allows inode reclaim 148 * @mapping: the page cache mapping 173 * inodes before there is highmem pressure from the page 186 * head pointer, which allows non-resident page cache entries 317 * reduce page locking. 425 struct address_space *page_mapping(struct page *); 433 * For folios which are in the page cache, return the mapping that this 434 * page belongs to. Folios in the swap cache return the mapping of the 439 * Do not call this for folios which aren't in the page cache or swap cache. 453 * For folios which are in the page cach 469 page_file_mapping(struct page *page) argument 544 attach_page_private(struct page *page, void *data) argument 549 detach_page_private(struct page *page) argument 908 page_to_pgoff(struct page *page) argument 926 page_offset(struct page *page) argument 931 page_file_offset(struct page *page) argument 1025 trylock_page(struct page *page) argument 1070 lock_page(struct page *page) argument 1141 wait_on_page_locked(struct page *page) argument 1386 struct page *page; local 1399 VM_BUG_ON_PAGE(PageTail(page), page); local 1511 page_mkwrite_check_truncate(struct page *page, struct inode *inode) argument 1549 i_blocks_per_page(struct inode *inode, struct page *page) argument [all...] |
/linux-master/kernel/power/ |
H A D | snapshot.c | 83 * The calls to set_direct_map_*() should not fail because remapping a page 88 static inline void hibernate_map_page(struct page *page) argument 91 int ret = set_direct_map_default_noflush(page); 94 pr_warn_once("Failed to remap page\n"); 96 debug_pagealloc_map_pages(page, 1); 100 static inline void hibernate_unmap_page(struct page *page) argument 103 unsigned long addr = (unsigned long)page_address(page); 104 int ret = set_direct_map_invalid_noflush(page); 229 struct page *page; local 257 struct page *page; local 1042 swsusp_set_page_free(struct page *page) argument 1048 swsusp_page_is_free(struct page *page) argument 1054 swsusp_unset_page_free(struct page *page) argument 1060 swsusp_set_page_forbidden(struct page *page) argument 1066 swsusp_page_is_forbidden(struct page *page) argument 1072 swsusp_unset_page_forbidden(struct page *page) argument 1190 clear_or_poison_free_page(struct page *page) argument 1253 struct page *page; local 1323 struct page *page; local 1387 struct page *page; local 1626 struct page *page = pfn_to_page(fr_pfn); local 1662 struct page *page; local 1762 struct page *page = pfn_to_page(pfn); local 2064 struct page *page; local 2103 struct page *page; local 2270 struct page *page; local 2478 struct page *page; local 2515 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument 2598 get_highmem_page_buffer(struct page *page, struct chain_allocator *ca) argument 2729 struct page *page; local [all...] |
/linux-master/drivers/s390/char/ |
H A D | sclp_con.c | 50 void *page; local 53 page = sclp_unmake_buffer(buffer); 58 list_add_tail((struct list_head *) page, &sclp_con_pages); 139 void *page; local 151 page = sclp_unmake_buffer(buffer); 152 list_add_tail((struct list_head *) page, &sclp_con_pages); 164 void *page; local 186 page = sclp_con_pages.next; 187 list_del((struct list_head *) page); 188 sclp_conbuf = sclp_make_buffer(page, SCLP_CON_COLUMN 272 void *page; local [all...] |
/linux-master/arch/m68k/include/asm/ |
H A D | cacheflush_mm.h | 222 /* Push the page at kernel virtual address and clear the icache */ 258 #define flush_dcache_page(page) __flush_pages_to_ram(page_address(page), 1) 263 #define flush_icache_pages(vma, page, nr) \ 264 __flush_pages_to_ram(page_address(page), nr) 266 extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page, 273 struct page *page, unsigned long vaddr, 276 flush_cache_page(vma, vaddr, page_to_pfn(page)); 272 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument 280 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument [all...] |
/linux-master/arch/xtensa/include/asm/ |
H A D | page.h | 2 * include/asm-xtensa/page.h 22 * PAGE_SHIFT determines the page size 43 * If the cache size for one way is greater than the page size, we have to 44 * deal with cache aliasing. The cache index is wider than the page size: 55 * When the page number is translated to the physical page address, the lowest 59 * The kernel does not provide a mechanism to ensure that the page color 62 * the page might also change. 99 typedef struct { unsigned long pte; } pte_t; /* page table entry */ 102 typedef struct page *pgtable_ [all...] |
/linux-master/tools/virtio/linux/ |
H A D | scatterlist.h | 21 * sg_assign_page - Assign a given page to an SG entry 23 * @page: The page 26 * Assign page to sg entry. Also see sg_set_page(), the most commonly used 30 static inline void sg_assign_page(struct scatterlist *sg, struct page *page) argument 38 BUG_ON((unsigned long) page & 0x03); 42 sg->page_link = page_link | (unsigned long) page; 46 * sg_set_page - Set sg entry to point at given page 48 * @page 59 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len, unsigned int offset) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 16 struct page *page; local 18 for_each_sgt_page(page, sgt_iter, pages) { 19 __free_page(page); 52 struct page *page; local 54 page = alloc_page(GFP | __GFP_HIGHMEM); 55 if (!page) { 60 sg_set_page(sg, page, PAGE_SIZE, 0);
|
/linux-master/drivers/hwmon/pmbus/ |
H A D | max31785.c | 82 int page, int command) 87 rc = pmbus_read_byte_data(client, page, command); 94 int page, int command, u16 data) 99 rc = pmbus_write_byte_data(client, page, command, data); 106 int page, int phase, int command) 111 rc = pmbus_read_word_data(client, page, phase, command); 118 int page, int command, u16 data) 123 rc = pmbus_write_word_data(client, page, command, data); 128 static int max31785_read_byte_data(struct i2c_client *client, int page, int reg) argument 138 page 80 _max31785_read_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command) argument 92 _max31785_write_byte_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument 104 _max31785_read_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int phase, int command) argument 116 _max31785_write_word_data(struct i2c_client *client, struct max31785_data *driver_data, int page, int command, u16 data) argument 145 max31785_write_byte(struct i2c_client *client, int page, u8 value) argument 153 max31785_read_long_data(struct i2c_client *client, int page, int reg, u32 *data) argument 191 max31785_get_pwm(struct i2c_client *client, int page) argument 206 max31785_get_pwm_mode(struct i2c_client *client, struct max31785_data *driver_data, int page) argument 233 max31785_read_word_data(struct i2c_client *client, int page, int phase, int reg) argument 297 max31785_update_fan(struct i2c_client *client, struct max31785_data *driver_data, int page, u8 config, u8 mask, u16 command) argument 324 max31785_pwm_enable(struct i2c_client *client, struct max31785_data *driver_data, int page, u16 word) argument 358 max31785_write_word_data(struct i2c_client *client, int page, int reg, u16 word) argument [all...] |
/linux-master/drivers/gpu/drm/gma500/ |
H A D | mmu.h | 13 * before taking the page table spinlock. 17 /* protects page tables, directory tables and pt tables. 39 struct page *p; 47 struct page *p; 48 struct page *dummy_pt; 49 struct page *dummy_page; 77 extern int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages,
|
/linux-master/drivers/usb/gadget/function/ |
H A D | u_ether_configfs.h | 30 char *page) \ 36 result = gether_get_dev_addr(opts->net, page, PAGE_SIZE); \ 43 const char *page, size_t len)\ 54 ret = gether_set_dev_addr(opts->net, page); \ 65 char *page) \ 71 result = gether_get_host_addr(opts->net, page, PAGE_SIZE); \ 78 const char *page, size_t len)\ 89 ret = gether_set_host_addr(opts->net, page); \ 100 char *page) \ 108 return sprintf(page, " [all...] |
/linux-master/tools/testing/nvdimm/ |
H A D | pmem-dax.c | 22 * Limit dax to a single page at a time given vmalloc()-backed 26 struct page *page; local 30 page = vmalloc_to_page(pmem->virt_addr + offset); 32 *pfn = page_to_pfn_t(page); 34 __func__, pmem, pgoff, page_to_pfn(page));
|
/linux-master/arch/arm/mm/ |
H A D | copypage-xscale.c | 8 * processors. When we copy a user page page, we map it in such a way 9 * that accesses to this page will not touch the main data cache, but 11 * the main data cache on page faults. 84 void xscale_mc_copy_user_highpage(struct page *to, struct page *from, 108 xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 110 void *ptr, *kaddr = kmap_atomic(page);
|
H A D | copypage-v4mc.c | 8 * processors. When we copy a user page page, we map it in such a way 9 * that accesses to this page will not touch the main data cache, but 11 * the main data cache on page faults. 64 void v4_mc_copy_user_highpage(struct page *to, struct page *from, 87 void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 89 void *ptr, *kaddr = kmap_atomic(page);
|
/linux-master/arch/csky/abiv2/inc/abi/ |
H A D | cacheflush.h | 29 static inline void flush_dcache_page(struct page *page) argument 31 flush_dcache_folio(page_folio(page)); 47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 58 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/arch/x86/include/asm/ |
H A D | page.h | 19 struct page; 25 static inline void clear_user_page(void *page, unsigned long vaddr, argument 26 struct page *pg) 28 clear_page(page); 32 struct page *topage)
|
/linux-master/fs/f2fs/ |
H A D | dir.c | 166 struct page *dentry_page, 303 struct page **res_page) 308 struct page *dentry_page; 362 struct page **res_page) 401 * It returns the page where the entry was found (as a parameter - res_page), 406 const struct qstr *child, struct page **res_page) 427 struct f2fs_dir_entry *f2fs_parent_dir(struct inode *dir, struct page **p) 433 struct page **page) 438 de = f2fs_find_entry(dir, qstr, page); 432 f2fs_inode_by_name(struct inode *dir, const struct qstr *qstr, struct page **page) argument 447 f2fs_set_link(struct inode *dir, struct f2fs_dir_entry *de, struct page *page, struct inode *inode) argument 512 make_empty_dir(struct inode *inode, struct inode *parent, struct page *page) argument 539 struct page *page; local 689 struct page *page = NULL; local 802 struct page *page = NULL; local 836 struct page *page; local 881 f2fs_delete_entry(struct f2fs_dir_entry *dentry, struct page *page, struct inode *dir, struct inode *inode) argument [all...] |
H A D | recovery.c | 167 static int recover_dentry(struct inode *inode, struct page *ipage, 175 struct page *page; local 197 de = __f2fs_find_entry(dir, &fname, &page); 222 f2fs_delete_entry(de, page, dir, einode); 225 } else if (IS_ERR(page)) { 226 err = PTR_ERR(page); 236 f2fs_put_page(page, 0); 248 static int recover_quota_data(struct inode *inode, struct page *page) argument 287 recover_inode(struct inode *inode, struct page *page) argument 367 struct page *page = NULL; local 410 struct page *page = NULL; local 626 do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page) argument 782 struct page *page = NULL; local [all...] |
/linux-master/scripts/kconfig/lxdialog/ |
H A D | textbox.c | 13 static const char *buf, *page; variable 18 * 'page' will be updated to point to the desired line in 'buf'. 27 if (*page == '\0') { 33 if (page == buf) { 37 page--; 39 if (page == buf) { 43 page--; 44 } while (*page != '\n'); 45 page++; 51 * 'page' shoul [all...] |
/linux-master/arch/sh/mm/ |
H A D | nommu.c | 16 #include <asm/page.h> 55 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 60 void local_flush_tlb_one(unsigned long asid, unsigned long page) argument 82 void *kmap_coherent(struct page *page, unsigned long addr) argument
|
/linux-master/drivers/net/dsa/b53/ |
H A D | b53_spi.c | 76 static inline int b53_spi_set_page(struct spi_device *spi, u8 page) argument 82 txbuf[2] = page; 87 static inline int b53_prepare_reg_access(struct spi_device *spi, u8 page) argument 94 return b53_spi_set_page(spi, page); 124 static int b53_spi_read(struct b53_device *dev, u8 page, u8 reg, u8 *data, argument 130 ret = b53_prepare_reg_access(spi, page); 141 static int b53_spi_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument 143 return b53_spi_read(dev, page, reg, val, 1); 146 static int b53_spi_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument 151 ret = b53_spi_read(dev, page, re 159 b53_spi_read32(struct b53_device *dev, u8 page, u8 reg, u32 *val) argument 172 b53_spi_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 185 b53_spi_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 198 b53_spi_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument 215 b53_spi_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument 232 b53_spi_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument 249 b53_spi_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument 266 b53_spi_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument [all...] |
/linux-master/drivers/char/agp/ |
H A D | efficeon-agp.c | 28 #include <linux/page-flags.h> 38 * The page array is filled through the ATTPAGE register 40 * 31:20: physical page address 43 * mapped pages (the 2nd level page table pages 69 static inline unsigned long efficeon_mask_memory(struct page *page) argument 71 unsigned long addr = page_to_phys(page); 163 unsigned long page = efficeon_private.l1_table[index]; local 164 if (page) { 166 free_page(page); 210 unsigned long page; local 239 unsigned int *page, *last_page; local 300 unsigned int *page = (unsigned int *) efficeon_private.l1_table[index >> 10]; local [all...] |