Searched refs:page (Results 251 - 275 of 3144) sorted by relevance

<<11121314151617181920>>

/linux-master/fs/f2fs/
H A Drecovery.c167 static int recover_dentry(struct inode *inode, struct page *ipage,
175 struct page *page; local
197 de = __f2fs_find_entry(dir, &fname, &page);
222 f2fs_delete_entry(de, page, dir, einode);
225 } else if (IS_ERR(page)) {
226 err = PTR_ERR(page);
236 f2fs_put_page(page, 0);
248 static int recover_quota_data(struct inode *inode, struct page *page) argument
287 recover_inode(struct inode *inode, struct page *page) argument
368 struct page *page = NULL; local
411 struct page *page = NULL; local
627 do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page) argument
783 struct page *page = NULL; local
[all...]
H A Ddir.c166 struct page *dentry_page,
303 struct page **res_page)
308 struct page *dentry_page;
362 struct page **res_page)
401 * It returns the page where the entry was found (as a parameter - res_page),
406 const struct qstr *child, struct page **res_page)
427 struct f2fs_dir_entry *f2fs_parent_dir(struct inode *dir, struct page **p)
433 struct page **page)
438 de = f2fs_find_entry(dir, qstr, page);
432 f2fs_inode_by_name(struct inode *dir, const struct qstr *qstr, struct page **page) argument
447 f2fs_set_link(struct inode *dir, struct f2fs_dir_entry *de, struct page *page, struct inode *inode) argument
512 make_empty_dir(struct inode *inode, struct inode *parent, struct page *page) argument
539 struct page *page; local
689 struct page *page = NULL; local
802 struct page *page = NULL; local
836 struct page *page; local
881 f2fs_delete_entry(struct f2fs_dir_entry *dentry, struct page *page, struct inode *dir, struct inode *inode) argument
[all...]
H A Dnode.c107 * free memory is lower than watermark or cached page count
108 * exceed threshold, deny caching compress page.
123 static void clear_node_page_dirty(struct page *page) argument
125 if (PageDirty(page)) {
126 f2fs_clear_page_cache_dirty_tag(page);
127 clear_page_dirty_for_io(page);
128 dec_page_count(F2FS_P_SB(page), F2FS_DIRTY_NODES);
130 ClearPageUptodate(page);
133 static struct page *get_current_nat_pag
313 f2fs_in_warm_node_list(struct f2fs_sb_info *sbi, struct page *page) argument
327 f2fs_add_fsync_node_entry(struct f2fs_sb_info *sbi, struct page *page) argument
351 f2fs_del_fsync_node_entry(struct f2fs_sb_info *sbi, struct page *page) argument
554 struct page *page = NULL; local
937 struct page *page; local
976 struct page *page; local
1122 struct page *page; local
1301 struct page *page; local
1360 read_node_page(struct page *page, blk_opf_t op_flags) argument
1430 struct page *page; local
1506 struct page *page; local
1555 struct page *page = &fbatch.folios[i]->page; local
1596 __write_node_page(struct page *page, bool atomic, bool *submitted, struct writeback_control *wbc, bool do_balance, enum iostat_type io_type, unsigned int *seq_id) argument
1746 f2fs_write_node_page(struct page *page, struct writeback_control *wbc) argument
1781 struct page *page = &fbatch.folios[i]->page; local
1897 flush_dirty_inode(struct page *page) argument
1928 struct page *page = &fbatch.folios[i]->page; local
1982 struct page *page = &fbatch.folios[i]->page; local
2086 struct page *page; local
2505 struct page *page = get_current_nat_page(sbi, nid); local
2687 f2fs_recover_inline_xattr(struct inode *inode, struct page *page) argument
2724 f2fs_recover_xattr_data(struct inode *inode, struct page *page) argument
2772 f2fs_recover_inode_page(struct f2fs_sb_info *sbi, struct page *page) argument
2863 struct page *page = f2fs_get_tmp_page(sbi, idx); local
2958 update_nat_bits(struct f2fs_sb_info *sbi, nid_t start_nid, struct page *page) argument
3019 struct page *page = NULL; local
3172 struct page *page; local
[all...]
/linux-master/scripts/kconfig/lxdialog/
H A Dtextbox.c13 static const char *buf, *page; variable
18 * 'page' will be updated to point to the desired line in 'buf'.
27 if (*page == '\0') {
33 if (page == buf) {
37 page--;
39 if (page == buf) {
43 page--;
44 } while (*page != '\n');
45 page++;
51 * 'page' shoul
[all...]
/linux-master/drivers/net/dsa/b53/
H A Db53_spi.c76 static inline int b53_spi_set_page(struct spi_device *spi, u8 page) argument
82 txbuf[2] = page;
87 static inline int b53_prepare_reg_access(struct spi_device *spi, u8 page) argument
94 return b53_spi_set_page(spi, page);
124 static int b53_spi_read(struct b53_device *dev, u8 page, u8 reg, u8 *data, argument
130 ret = b53_prepare_reg_access(spi, page);
141 static int b53_spi_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument
143 return b53_spi_read(dev, page, reg, val, 1);
146 static int b53_spi_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument
151 ret = b53_spi_read(dev, page, re
159 b53_spi_read32(struct b53_device *dev, u8 page, u8 reg, u32 *val) argument
172 b53_spi_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument
185 b53_spi_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument
198 b53_spi_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument
215 b53_spi_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument
232 b53_spi_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument
249 b53_spi_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument
266 b53_spi_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument
[all...]
H A Db53_mdio.c42 static int b53_mdio_op(struct b53_device *dev, u8 page, u8 reg, u16 op) argument
49 if (dev->current_page != page) {
50 /* set page number */
51 v = (page << 8) | REG_MII_PAGE_ENABLE;
56 dev->current_page = page;
80 static int b53_mdio_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument
85 ret = b53_mdio_op(dev, page, reg, REG_MII_ADDR_READ);
95 static int b53_mdio_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument
100 ret = b53_mdio_op(dev, page, reg, REG_MII_ADDR_READ);
109 static int b53_mdio_read32(struct b53_device *dev, u8 page, u argument
125 b53_mdio_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument
147 b53_mdio_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument
169 b53_mdio_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument
182 b53_mdio_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument
196 b53_mdio_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument
215 b53_mdio_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument
234 b53_mdio_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument
[all...]
/linux-master/drivers/char/agp/
H A Defficeon-agp.c28 #include <linux/page-flags.h>
38 * The page array is filled through the ATTPAGE register
40 * 31:20: physical page address
43 * mapped pages (the 2nd level page table pages
69 static inline unsigned long efficeon_mask_memory(struct page *page) argument
71 unsigned long addr = page_to_phys(page);
163 unsigned long page = efficeon_private.l1_table[index]; local
164 if (page) {
166 free_page(page);
210 unsigned long page; local
239 unsigned int *page, *last_page; local
300 unsigned int *page = (unsigned int *) efficeon_private.l1_table[index >> 10]; local
[all...]
/linux-master/mm/
H A Ddmapool.c10 * the given device. It uses the dma_alloc_coherent page allocator to get
16 * allocated pages. Each page in the page_list is split into blocks of at
19 * keep a count of how many are currently allocated from each page.
122 struct dma_page *page; local
124 list_for_each_entry(page, &pool->page_list, page_list) {
125 if (dma < page->dma)
127 if ((dma - page->dma) < pool->allocation)
128 return page;
136 struct dma_page *page; local
138 page
159 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument
176 pool_init_page(struct dma_pool *pool, struct dma_page *page) argument
300 pool_initialise_page(struct dma_pool *pool, struct dma_page *page) argument
336 struct dma_page *page; local
362 struct dma_page *page, *tmp; local
408 struct dma_page *page; local
[all...]
H A Dksm.c76 * by their contents. Because each such page is write-protected, searching on
81 * mapping from a KSM page to virtual addresses that map this page.
90 * different KSM page copy of that content
114 * take 10 attempts to find a page in the unstable tree, once it is found,
115 * it is secured in the stable tree. (When we scan a new page, we first
150 * @node: rb node of this ksm page in the stable tree
154 * @hlist: hlist head of rmap_items using this ksm page
155 * @kpfn: page frame number of this ksm page (perhap
618 struct page *page = NULL; local
774 struct page *page; local
921 struct page *page; local
1001 struct page *page; local
1092 page_stable_node(struct page *page) argument
1097 set_page_stable_node(struct page *page, struct ksm_stable_node *stable_node) argument
1110 struct page *page; local
1269 calc_checksum(struct page *page) argument
1278 write_protect_page(struct vm_area_struct *vma, struct page *page, pte_t *orig_pte) argument
1370 replace_page(struct vm_area_struct *vma, struct page *page, struct page *kpage, pte_t orig_pte) argument
1412 VM_BUG_ON_PAGE(PageAnonExclusive(page), page); local
1475 try_to_merge_one_page(struct vm_area_struct *vma, struct page *page, struct page *kpage) argument
1540 try_to_merge_with_ksm_page(struct ksm_rmap_item *rmap_item, struct page *page, struct page *kpage) argument
1579 try_to_merge_two_pages(struct ksm_rmap_item *rmap_item, struct page *page, struct ksm_rmap_item *tree_rmap_item, struct page *tree_page) argument
1817 stable_tree_search(struct page *page) argument
2187 unstable_tree_search_insert(struct ksm_rmap_item *rmap_item, struct page *page, struct page **tree_pagep) argument
2301 cmp_and_merge_page(struct page *page, struct ksm_rmap_item *rmap_item) argument
2519 should_skip_rmap_item(struct page *page, struct ksm_rmap_item *rmap_item) argument
2562 scan_get_next_rmap_item(struct page **page) argument
2600 struct page *page; local
2749 struct page *page; local
3055 struct page *page = folio_page(folio, 0); local
3175 collect_procs_ksm(struct page *page, struct list_head *to_kill, int force_early) argument
[all...]
H A Dswap_cgroup.c10 struct page **map;
41 struct page *page; local
48 page = alloc_page(GFP_KERNEL | __GFP_ZERO);
49 if (!page)
51 ctrl->map[idx] = page;
68 struct page *mappage;
206 struct page **map;
223 struct page *page local
[all...]
/linux-master/fs/crypto/
H A Dcrypto.c50 struct page *fscrypt_alloc_bounce_page(gfp_t gfp_flags)
55 * page pool, but it didn't set needs_bounce_pages.
63 * fscrypt_free_bounce_page() - free a ciphertext bounce page
64 * @bounce_page: the bounce page to free, or NULL
66 * Free a bounce page that was allocated by fscrypt_encrypt_pagecache_blocks(),
69 void fscrypt_free_bounce_page(struct page *bounce_page)
110 struct page *src_page, struct page *dest_page,
156 * fscrypt_encrypt_pagecache_blocks() - Encrypt data from a pagecache page
157 * @page
180 fscrypt_encrypt_pagecache_blocks(struct page *page, unsigned int len, unsigned int offs, gfp_t gfp_flags) argument
240 fscrypt_encrypt_block_inplace(const struct inode *inode, struct page *page, unsigned int len, unsigned int offs, u64 lblk_num, gfp_t gfp_flags) argument
285 struct page *page = folio_page(folio, i >> PAGE_SHIFT); local
315 fscrypt_decrypt_block_inplace(const struct inode *inode, struct page *page, unsigned int len, unsigned int offs, u64 lblk_num) argument
[all...]
/linux-master/arch/mips/mm/
H A Dinit.c51 * any price. Since page is never written to after the initialization we
64 struct page *page; local
75 page = virt_to_page((void *)empty_zero_page);
76 split_page(page, order);
77 for (i = 0; i < (1 << order); i++, page++)
78 mark_page_reserved(page);
83 static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot) argument
92 BUG_ON(folio_test_dcache_dirty(page_folio(page)));
138 kmap_coherent(struct page *page, unsigned long addr) argument
143 kmap_noncoherent(struct page *page, unsigned long addr) argument
195 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument
215 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument
453 struct page *page = pfn_to_page(tmp); local
491 struct page *page = pfn_to_page(pfn); local
[all...]
/linux-master/drivers/hwmon/pmbus/
H A Dpmbus.c26 * Find sensor groups and status registers on each page.
31 int page; local
33 /* Sensors detected on page 0 only */
70 for (page = 0; page < info->pages; page++) {
71 if (pmbus_check_word_register(client, page, PMBUS_READ_VOUT)) {
72 info->func[page] |= PMBUS_HAVE_VOUT;
73 if (pmbus_check_byte_register(client, page,
75 info->func[page] |
104 int page; local
[all...]
H A Dmax15301.c69 static int max15301_read_word_data(struct i2c_client *client, int page, argument
76 if (page > 0)
83 ret = pmbus_read_word_data(client, page, phase, reg);
89 static int max15301_read_byte_data(struct i2c_client *client, int page, int reg) argument
95 if (page > 0)
99 ret = pmbus_read_byte_data(client, page, reg);
105 static int max15301_write_word_data(struct i2c_client *client, int page, int reg, argument
112 if (page > 0)
119 ret = pmbus_write_word_data(client, page, reg, word);
125 static int max15301_write_byte(struct i2c_client *client, int page, u argument
[all...]
H A Dmax16064.c18 static int max16064_read_word_data(struct i2c_client *client, int page, argument
25 ret = pmbus_read_word_data(client, page, phase,
29 ret = pmbus_read_word_data(client, page, phase,
43 static int max16064_write_word_data(struct i2c_client *client, int page, argument
50 ret = pmbus_write_word_data(client, page,
54 ret = pmbus_write_word_data(client, page,
/linux-master/arch/mips/include/asm/
H A Dtlbflush.h12 * - flush_tlb_page(vma, vmaddr) flushes one page
22 unsigned long page);
44 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
/linux-master/tools/testing/scatterlist/linux/
H A Dmm.h45 static inline unsigned long page_to_phys(struct page *page) argument
52 #define page_to_pfn(page) ((unsigned long)(page) / PAGE_SIZE)
54 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n))
79 static inline void *kmap(struct page *page) argument
86 static inline void *kmap_atomic(struct page *page) argument
108 free_page(unsigned long page) argument
[all...]
/linux-master/arch/loongarch/include/asm/
H A Dtlbflush.h15 * - flush_tlb_page(vma, vmaddr) flushes one page
25 extern void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
43 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
/linux-master/drivers/net/thunderbolt/
H A Dtrace.h26 TP_PROTO(unsigned int index, const void *page, dma_addr_t phys,
28 TP_ARGS(index, page, phys, dir),
31 __field(const void *, page)
37 __entry->page = page;
41 TP_printk("index=%u page=%p phys=%pad dir=%s",
42 __entry->index, __entry->page, &__entry->phys,
47 TP_PROTO(unsigned int index, const void *page, dma_addr_t phys,
49 TP_ARGS(index, page, phys, dir)
53 TP_PROTO(unsigned int index, const void *page, dma_addr_
[all...]
/linux-master/arch/hexagon/include/asm/
H A Dcacheflush.h17 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page
20 * - flush_dcache_page(pg) flushes(wback&invalidates) a page for dcache
71 void copy_to_user_page(struct vm_area_struct *vma, struct page *page,
75 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/linux-master/arch/arc/mm/
H A Ddma.c18 void arch_dma_prep_coherent(struct page *page, size_t size) argument
21 * Evict any existing L1 and/or L2 lines for the backing page
22 * in case it was used earlier as a normal "cached" page.
30 dma_cache_wback_inv(page_to_phys(page), size);
/linux-master/arch/riscv/include/asm/
H A Dcacheflush.h26 static inline void flush_dcache_page(struct page *page) argument
28 flush_dcache_folio(page_folio(page));
/linux-master/arch/arm64/mm/
H A Dpageattr.c27 * mapped at page granularity, so that it is possible to
30 * KFENCE pool requires page-granular mapping if initialized late.
84 * mappings into page mappings may cause TLB conflicts. This means
89 * Those are guaranteed to consist entirely of page mappings, and
165 int set_direct_map_invalid_noflush(struct page *page) argument
176 (unsigned long)page_address(page),
180 int set_direct_map_default_noflush(struct page *page) argument
191 (unsigned long)page_address(page),
196 __kernel_map_pages(struct page *page, int numpages, int enable) argument
213 kernel_page_present(struct page *page) argument
[all...]
/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_memfree.c40 #include <asm/page.h>
61 } page[]; member in struct:mthca_user_db_table
109 struct page *page; local
115 page = alloc_pages(gfp_mask | __GFP_ZERO, order);
116 if (!page)
119 sg_set_page(mem, page, PAGE_SIZE << order, 0);
283 struct page *page = NULL; local
306 * so if we found the page, dma_handl
439 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page) argument
575 struct mthca_db_page *page; local
670 struct mthca_db_page *page; local
[all...]
/linux-master/include/linux/
H A Dgfp.h142 * There is only one page-allocator function, and two main namespaces to
143 * it. The alloc_page*() variants return 'struct page *' and as such
144 * can allocate highmem pages, the *get*page*() variants return
145 * virtual kernel addresses to the allocated page(s).
172 static inline void arch_free_page(struct page *page, int order) { } argument
175 static inline void arch_alloc_page(struct page *page, int order) { } argument
178 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid,
186 struct page **page_arra
[all...]

Completed in 312 milliseconds

<<11121314151617181920>>