/linux-master/fs/f2fs/ |
H A D | recovery.c | 167 static int recover_dentry(struct inode *inode, struct page *ipage, 175 struct page *page; local 197 de = __f2fs_find_entry(dir, &fname, &page); 222 f2fs_delete_entry(de, page, dir, einode); 225 } else if (IS_ERR(page)) { 226 err = PTR_ERR(page); 236 f2fs_put_page(page, 0); 248 static int recover_quota_data(struct inode *inode, struct page *page) argument 287 recover_inode(struct inode *inode, struct page *page) argument 368 struct page *page = NULL; local 411 struct page *page = NULL; local 627 do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page) argument 783 struct page *page = NULL; local [all...] |
H A D | dir.c | 166 struct page *dentry_page, 303 struct page **res_page) 308 struct page *dentry_page; 362 struct page **res_page) 401 * It returns the page where the entry was found (as a parameter - res_page), 406 const struct qstr *child, struct page **res_page) 427 struct f2fs_dir_entry *f2fs_parent_dir(struct inode *dir, struct page **p) 433 struct page **page) 438 de = f2fs_find_entry(dir, qstr, page); 432 f2fs_inode_by_name(struct inode *dir, const struct qstr *qstr, struct page **page) argument 447 f2fs_set_link(struct inode *dir, struct f2fs_dir_entry *de, struct page *page, struct inode *inode) argument 512 make_empty_dir(struct inode *inode, struct inode *parent, struct page *page) argument 539 struct page *page; local 689 struct page *page = NULL; local 802 struct page *page = NULL; local 836 struct page *page; local 881 f2fs_delete_entry(struct f2fs_dir_entry *dentry, struct page *page, struct inode *dir, struct inode *inode) argument [all...] |
H A D | node.c | 107 * free memory is lower than watermark or cached page count 108 * exceed threshold, deny caching compress page. 123 static void clear_node_page_dirty(struct page *page) argument 125 if (PageDirty(page)) { 126 f2fs_clear_page_cache_dirty_tag(page); 127 clear_page_dirty_for_io(page); 128 dec_page_count(F2FS_P_SB(page), F2FS_DIRTY_NODES); 130 ClearPageUptodate(page); 133 static struct page *get_current_nat_pag 313 f2fs_in_warm_node_list(struct f2fs_sb_info *sbi, struct page *page) argument 327 f2fs_add_fsync_node_entry(struct f2fs_sb_info *sbi, struct page *page) argument 351 f2fs_del_fsync_node_entry(struct f2fs_sb_info *sbi, struct page *page) argument 554 struct page *page = NULL; local 937 struct page *page; local 976 struct page *page; local 1122 struct page *page; local 1301 struct page *page; local 1360 read_node_page(struct page *page, blk_opf_t op_flags) argument 1430 struct page *page; local 1506 struct page *page; local 1555 struct page *page = &fbatch.folios[i]->page; local 1596 __write_node_page(struct page *page, bool atomic, bool *submitted, struct writeback_control *wbc, bool do_balance, enum iostat_type io_type, unsigned int *seq_id) argument 1746 f2fs_write_node_page(struct page *page, struct writeback_control *wbc) argument 1781 struct page *page = &fbatch.folios[i]->page; local 1897 flush_dirty_inode(struct page *page) argument 1928 struct page *page = &fbatch.folios[i]->page; local 1982 struct page *page = &fbatch.folios[i]->page; local 2086 struct page *page; local 2505 struct page *page = get_current_nat_page(sbi, nid); local 2687 f2fs_recover_inline_xattr(struct inode *inode, struct page *page) argument 2724 f2fs_recover_xattr_data(struct inode *inode, struct page *page) argument 2772 f2fs_recover_inode_page(struct f2fs_sb_info *sbi, struct page *page) argument 2863 struct page *page = f2fs_get_tmp_page(sbi, idx); local 2958 update_nat_bits(struct f2fs_sb_info *sbi, nid_t start_nid, struct page *page) argument 3019 struct page *page = NULL; local 3172 struct page *page; local [all...] |
/linux-master/scripts/kconfig/lxdialog/ |
H A D | textbox.c | 13 static const char *buf, *page; variable 18 * 'page' will be updated to point to the desired line in 'buf'. 27 if (*page == '\0') { 33 if (page == buf) { 37 page--; 39 if (page == buf) { 43 page--; 44 } while (*page != '\n'); 45 page++; 51 * 'page' shoul [all...] |
/linux-master/drivers/net/dsa/b53/ |
H A D | b53_spi.c | 76 static inline int b53_spi_set_page(struct spi_device *spi, u8 page) argument 82 txbuf[2] = page; 87 static inline int b53_prepare_reg_access(struct spi_device *spi, u8 page) argument 94 return b53_spi_set_page(spi, page); 124 static int b53_spi_read(struct b53_device *dev, u8 page, u8 reg, u8 *data, argument 130 ret = b53_prepare_reg_access(spi, page); 141 static int b53_spi_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument 143 return b53_spi_read(dev, page, reg, val, 1); 146 static int b53_spi_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument 151 ret = b53_spi_read(dev, page, re 159 b53_spi_read32(struct b53_device *dev, u8 page, u8 reg, u32 *val) argument 172 b53_spi_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 185 b53_spi_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 198 b53_spi_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument 215 b53_spi_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument 232 b53_spi_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument 249 b53_spi_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument 266 b53_spi_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument [all...] |
H A D | b53_mdio.c | 42 static int b53_mdio_op(struct b53_device *dev, u8 page, u8 reg, u16 op) argument 49 if (dev->current_page != page) { 50 /* set page number */ 51 v = (page << 8) | REG_MII_PAGE_ENABLE; 56 dev->current_page = page; 80 static int b53_mdio_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument 85 ret = b53_mdio_op(dev, page, reg, REG_MII_ADDR_READ); 95 static int b53_mdio_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument 100 ret = b53_mdio_op(dev, page, reg, REG_MII_ADDR_READ); 109 static int b53_mdio_read32(struct b53_device *dev, u8 page, u argument 125 b53_mdio_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 147 b53_mdio_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 169 b53_mdio_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument 182 b53_mdio_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument 196 b53_mdio_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument 215 b53_mdio_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument 234 b53_mdio_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument [all...] |
/linux-master/drivers/char/agp/ |
H A D | efficeon-agp.c | 28 #include <linux/page-flags.h> 38 * The page array is filled through the ATTPAGE register 40 * 31:20: physical page address 43 * mapped pages (the 2nd level page table pages 69 static inline unsigned long efficeon_mask_memory(struct page *page) argument 71 unsigned long addr = page_to_phys(page); 163 unsigned long page = efficeon_private.l1_table[index]; local 164 if (page) { 166 free_page(page); 210 unsigned long page; local 239 unsigned int *page, *last_page; local 300 unsigned int *page = (unsigned int *) efficeon_private.l1_table[index >> 10]; local [all...] |
/linux-master/mm/ |
H A D | dmapool.c | 10 * the given device. It uses the dma_alloc_coherent page allocator to get 16 * allocated pages. Each page in the page_list is split into blocks of at 19 * keep a count of how many are currently allocated from each page. 122 struct dma_page *page; local 124 list_for_each_entry(page, &pool->page_list, page_list) { 125 if (dma < page->dma) 127 if ((dma - page->dma) < pool->allocation) 128 return page; 136 struct dma_page *page; local 138 page 159 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument 176 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument 300 pool_initialise_page(struct dma_pool *pool, struct dma_page *page) argument 336 struct dma_page *page; local 362 struct dma_page *page, *tmp; local 408 struct dma_page *page; local [all...] |
H A D | ksm.c | 76 * by their contents. Because each such page is write-protected, searching on 81 * mapping from a KSM page to virtual addresses that map this page. 90 * different KSM page copy of that content 114 * take 10 attempts to find a page in the unstable tree, once it is found, 115 * it is secured in the stable tree. (When we scan a new page, we first 150 * @node: rb node of this ksm page in the stable tree 154 * @hlist: hlist head of rmap_items using this ksm page 155 * @kpfn: page frame number of this ksm page (perhap 618 struct page *page = NULL; local 774 struct page *page; local 921 struct page *page; local 1001 struct page *page; local 1092 page_stable_node(struct page *page) argument 1097 set_page_stable_node(struct page *page, struct ksm_stable_node *stable_node) argument 1110 struct page *page; local 1269 calc_checksum(struct page *page) argument 1278 write_protect_page(struct vm_area_struct *vma, struct page *page, pte_t *orig_pte) argument 1370 replace_page(struct vm_area_struct *vma, struct page *page, struct page *kpage, pte_t orig_pte) argument 1412 VM_BUG_ON_PAGE(PageAnonExclusive(page), page); local 1475 try_to_merge_one_page(struct vm_area_struct *vma, struct page *page, struct page *kpage) argument 1540 try_to_merge_with_ksm_page(struct ksm_rmap_item *rmap_item, struct page *page, struct page *kpage) argument 1579 try_to_merge_two_pages(struct ksm_rmap_item *rmap_item, struct page *page, struct ksm_rmap_item *tree_rmap_item, struct page *tree_page) argument 1817 stable_tree_search(struct page *page) argument 2187 unstable_tree_search_insert(struct ksm_rmap_item *rmap_item, struct page *page, struct page **tree_pagep) argument 2301 cmp_and_merge_page(struct page *page, struct ksm_rmap_item *rmap_item) argument 2519 should_skip_rmap_item(struct page *page, struct ksm_rmap_item *rmap_item) argument 2562 scan_get_next_rmap_item(struct page **page) argument 2600 struct page *page; local 2749 struct page *page; local 3055 struct page *page = folio_page(folio, 0); local 3175 collect_procs_ksm(struct page *page, struct list_head *to_kill, int force_early) argument [all...] |
H A D | swap_cgroup.c | 10 struct page **map; 41 struct page *page; local 48 page = alloc_page(GFP_KERNEL | __GFP_ZERO); 49 if (!page) 51 ctrl->map[idx] = page; 68 struct page *mappage; 206 struct page **map; 223 struct page *page local [all...] |
/linux-master/fs/crypto/ |
H A D | crypto.c | 50 struct page *fscrypt_alloc_bounce_page(gfp_t gfp_flags) 55 * page pool, but it didn't set needs_bounce_pages. 63 * fscrypt_free_bounce_page() - free a ciphertext bounce page 64 * @bounce_page: the bounce page to free, or NULL 66 * Free a bounce page that was allocated by fscrypt_encrypt_pagecache_blocks(), 69 void fscrypt_free_bounce_page(struct page *bounce_page) 110 struct page *src_page, struct page *dest_page, 156 * fscrypt_encrypt_pagecache_blocks() - Encrypt data from a pagecache page 157 * @page 180 fscrypt_encrypt_pagecache_blocks(struct page *page, unsigned int len, unsigned int offs, gfp_t gfp_flags) argument 240 fscrypt_encrypt_block_inplace(const struct inode *inode, struct page *page, unsigned int len, unsigned int offs, u64 lblk_num, gfp_t gfp_flags) argument 285 struct page *page = folio_page(folio, i >> PAGE_SHIFT); local 315 fscrypt_decrypt_block_inplace(const struct inode *inode, struct page *page, unsigned int len, unsigned int offs, u64 lblk_num) argument [all...] |
/linux-master/arch/mips/mm/ |
H A D | init.c | 51 * any price. Since page is never written to after the initialization we 64 struct page *page; local 75 page = virt_to_page((void *)empty_zero_page); 76 split_page(page, order); 77 for (i = 0; i < (1 << order); i++, page++) 78 mark_page_reserved(page); 83 static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot) argument 92 BUG_ON(folio_test_dcache_dirty(page_folio(page))); 138 kmap_coherent(struct page *page, unsigned long addr) argument 143 kmap_noncoherent(struct page *page, unsigned long addr) argument 195 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 215 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 453 struct page *page = pfn_to_page(tmp); local 491 struct page *page = pfn_to_page(pfn); local [all...] |
/linux-master/drivers/hwmon/pmbus/ |
H A D | pmbus.c | 26 * Find sensor groups and status registers on each page. 31 int page; local 33 /* Sensors detected on page 0 only */ 70 for (page = 0; page < info->pages; page++) { 71 if (pmbus_check_word_register(client, page, PMBUS_READ_VOUT)) { 72 info->func[page] |= PMBUS_HAVE_VOUT; 73 if (pmbus_check_byte_register(client, page, 75 info->func[page] | 104 int page; local [all...] |
H A D | max15301.c | 69 static int max15301_read_word_data(struct i2c_client *client, int page, argument 76 if (page > 0) 83 ret = pmbus_read_word_data(client, page, phase, reg); 89 static int max15301_read_byte_data(struct i2c_client *client, int page, int reg) argument 95 if (page > 0) 99 ret = pmbus_read_byte_data(client, page, reg); 105 static int max15301_write_word_data(struct i2c_client *client, int page, int reg, argument 112 if (page > 0) 119 ret = pmbus_write_word_data(client, page, reg, word); 125 static int max15301_write_byte(struct i2c_client *client, int page, u argument [all...] |
H A D | max16064.c | 18 static int max16064_read_word_data(struct i2c_client *client, int page, argument 25 ret = pmbus_read_word_data(client, page, phase, 29 ret = pmbus_read_word_data(client, page, phase, 43 static int max16064_write_word_data(struct i2c_client *client, int page, argument 50 ret = pmbus_write_word_data(client, page, 54 ret = pmbus_write_word_data(client, page,
|
/linux-master/arch/mips/include/asm/ |
H A D | tlbflush.h | 12 * - flush_tlb_page(vma, vmaddr) flushes one page 22 unsigned long page); 44 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/linux-master/tools/testing/scatterlist/linux/ |
H A D | mm.h | 45 static inline unsigned long page_to_phys(struct page *page) argument 52 #define page_to_pfn(page) ((unsigned long)(page) / PAGE_SIZE) 54 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) 79 static inline void *kmap(struct page *page) argument 86 static inline void *kmap_atomic(struct page *page) argument 108 free_page(unsigned long page) argument [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | tlbflush.h | 15 * - flush_tlb_page(vma, vmaddr) flushes one page 25 extern void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 43 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/linux-master/drivers/net/thunderbolt/ |
H A D | trace.h | 26 TP_PROTO(unsigned int index, const void *page, dma_addr_t phys, 28 TP_ARGS(index, page, phys, dir), 31 __field(const void *, page) 37 __entry->page = page; 41 TP_printk("index=%u page=%p phys=%pad dir=%s", 42 __entry->index, __entry->page, &__entry->phys, 47 TP_PROTO(unsigned int index, const void *page, dma_addr_t phys, 49 TP_ARGS(index, page, phys, dir) 53 TP_PROTO(unsigned int index, const void *page, dma_addr_ [all...] |
/linux-master/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 17 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page 20 * - flush_dcache_page(pg) flushes(wback&invalidates) a page for dcache 71 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, 75 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/arch/arc/mm/ |
H A D | dma.c | 18 void arch_dma_prep_coherent(struct page *page, size_t size) argument 21 * Evict any existing L1 and/or L2 lines for the backing page 22 * in case it was used earlier as a normal "cached" page. 30 dma_cache_wback_inv(page_to_phys(page), size);
|
/linux-master/arch/riscv/include/asm/ |
H A D | cacheflush.h | 26 static inline void flush_dcache_page(struct page *page) argument 28 flush_dcache_folio(page_folio(page));
|
/linux-master/arch/arm64/mm/ |
H A D | pageattr.c | 27 * mapped at page granularity, so that it is possible to 30 * KFENCE pool requires page-granular mapping if initialized late. 84 * mappings into page mappings may cause TLB conflicts. This means 89 * Those are guaranteed to consist entirely of page mappings, and 165 int set_direct_map_invalid_noflush(struct page *page) argument 176 (unsigned long)page_address(page), 180 int set_direct_map_default_noflush(struct page *page) argument 191 (unsigned long)page_address(page), 196 __kernel_map_pages(struct page *page, int numpages, int enable) argument 213 kernel_page_present(struct page *page) argument [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_memfree.c | 40 #include <asm/page.h> 61 } page[]; member in struct:mthca_user_db_table 109 struct page *page; local 115 page = alloc_pages(gfp_mask | __GFP_ZERO, order); 116 if (!page) 119 sg_set_page(mem, page, PAGE_SIZE << order, 0); 283 struct page *page = NULL; local 306 * so if we found the page, dma_handl 439 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page) argument 575 struct mthca_db_page *page; local 670 struct mthca_db_page *page; local [all...] |
/linux-master/include/linux/ |
H A D | gfp.h | 142 * There is only one page-allocator function, and two main namespaces to 143 * it. The alloc_page*() variants return 'struct page *' and as such 144 * can allocate highmem pages, the *get*page*() variants return 145 * virtual kernel addresses to the allocated page(s). 172 static inline void arch_free_page(struct page *page, int order) { } argument 175 static inline void arch_alloc_page(struct page *page, int order) { } argument 178 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, 186 struct page **page_arra [all...] |