Searched refs:folio (Results 51 - 75 of 361) sorted by relevance

1234567891011>>

/linux-master/mm/
H A Dhugetlb_vmemmap.h21 int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio);
25 void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio);
46 static inline int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio) argument
59 static inline void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio) argument
H A Ddebug.c54 static void __dump_folio(struct folio *folio, struct page *page, argument
57 struct address_space *mapping = folio_mapping(folio);
65 if (!folio_test_slab(folio)) {
67 if (folio_test_large(folio))
68 mapcount += folio_entire_mapcount(folio);
72 folio_ref_count(folio), mapcount, mapping,
73 folio->index + idx, pfn);
74 if (folio_test_large(folio)) {
76 folio_order(folio),
115 struct folio *foliop, folio; local
[all...]
H A Dinternal.h56 * If a 16GB hugetlb folio were mapped by PTEs of all of its 4kB pages,
72 * the folio's entire_mapcount.
74 static inline int folio_nr_pages_mapped(struct folio *folio) argument
76 return atomic_read(&folio->_nr_pages_mapped) & FOLIO_PAGES_MAPPED;
79 static inline void *folio_raw_mapping(struct folio *folio) argument
81 unsigned long mapping = (unsigned long)folio->mapping;
107 * folio_pte_batch - detect a PTE batch for a large folio
108 * @folio
129 folio_pte_batch(struct folio *folio, unsigned long addr, pte_t *start_ptep, pte_t pte, int max_nr, fpb_t flags, bool *any_writable) argument
144 VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio); local
181 acct_reclaim_writeback(struct folio *folio) argument
250 folio_evictable(struct folio *folio) argument
276 folio_needs_release(struct folio *folio) argument
499 folio_set_order(struct folio *folio, unsigned int order) argument
514 struct folio *folio = (struct folio *)page; local
522 struct folio *folio = (struct folio *)page; local
710 folio_within_range(struct folio *folio, struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
716 VM_WARN_ON_FOLIO(folio_test_ksm(folio), folio); local
738 folio_within_vma(struct folio *folio, struct vm_area_struct *vma) argument
753 mlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument
769 munlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument
872 unmap_mapping_folio(struct folio *folio) argument
873 mlock_new_folio(struct folio *folio) argument
[all...]
H A Dworkingset.c232 static void *lru_gen_eviction(struct folio *folio) argument
239 int type = folio_is_file_lru(folio);
240 int delta = folio_nr_pages(folio);
241 int refs = folio_lru_refs(folio);
243 struct mem_cgroup *memcg = folio_memcg(folio);
244 struct pglist_data *pgdat = folio_pgdat(folio);
260 * Tests if the shadow entry is for a folio that was recently evicted.
280 static void lru_gen_refault(struct folio *folio, voi argument
329 lru_gen_eviction(struct folio *folio) argument
340 lru_gen_refault(struct folio *folio, void *shadow) argument
382 workingset_eviction(struct folio *folio, struct mem_cgroup *target_memcg) argument
390 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local
391 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local
530 workingset_refault(struct folio *folio, void *shadow) argument
584 workingset_activation(struct folio *folio) argument
[all...]
H A Dmemory-failure.c610 static void collect_procs_anon(struct folio *folio, struct page *page, argument
618 av = folio_lock_anon_vma_read(folio, NULL);
647 static void collect_procs_file(struct folio *folio, struct page *page, argument
652 struct address_space *mapping = folio->mapping;
725 static void collect_procs(struct folio *folio, struct page *page, argument
728 if (!folio->mapping)
733 collect_procs_anon(folio, pag
923 delete_from_lru_cache(struct folio *folio) argument
948 truncate_error_folio(struct folio *folio, unsigned long pfn, struct address_space *mapping) argument
1034 struct folio *folio = page_folio(p); local
1158 struct folio *folio = page_folio(p); local
1180 struct folio *folio = page_folio(p); local
1202 struct folio *folio = page_folio(p); local
1404 struct folio *folio = page_folio(page); local
1501 struct folio *folio = page_folio(page); local
1573 struct folio *folio = page_folio(hpage); local
1748 struct folio *folio = pfn_folio(pfn); local
1859 raw_hwp_list_head(struct folio *folio) argument
1868 struct folio *folio = page_folio(page); local
1899 __folio_free_raw_hwp(struct folio *folio, bool move_flag) argument
1917 folio_set_hugetlb_hwpoison(struct folio *folio, struct page *page) argument
1960 folio_free_raw_hwp(struct folio *folio, bool move_flag) argument
1979 folio_clear_hugetlb_hwpoison(struct folio *folio) argument
2003 struct folio *folio = page_folio(page); local
2056 struct folio *folio; local
2125 folio_free_raw_hwp(struct folio *folio, bool flag) argument
2530 struct folio *folio; local
2631 mf_isolate_folio(struct folio *folio, struct list_head *pagelist) argument
2674 struct folio *folio = page_folio(page); local
[all...]
H A Dhugetlb.c59 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument
61 return cma_pages_valid(hugetlb_cma[folio_nid(folio)], &folio->page,
65 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument
1311 static void enqueue_hugetlb_folio(struct hstate *h, struct folio *folio) argument
1313 int nid = folio_nid(folio);
1316 VM_BUG_ON_FOLIO(folio_ref_count(folio), foli local
1327 struct folio *folio; local
1363 struct folio *folio; local
1395 struct folio *folio = NULL; local
1512 __destroy_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument
1535 destroy_compound_hugetlb_folio_for_demote(struct folio *folio, unsigned int order) argument
1542 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument
1548 free_gigantic_folio(struct folio *folio, unsigned int order) argument
1616 free_gigantic_folio(struct folio *folio, unsigned int order) argument
1618 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument
1622 __clear_hugetlb_destructor(struct hstate *h, struct folio *folio) argument
1640 __remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus, bool demote) argument
1646 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio(folio), folio); local
1647 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio_rsvd(folio), folio); local
1683 remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument
1689 remove_hugetlb_folio_for_demote(struct hstate *h, struct folio *folio, bool adjust_surplus) argument
1695 add_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument
1741 __update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument
1823 struct folio *folio; local
1851 update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio, bool atomic) argument
1874 struct folio *folio, *t_folio; local
1925 struct folio *folio, *t_folio; local
1973 free_huge_folio(struct folio *folio) argument
1985 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local
1986 VM_BUG_ON_FOLIO(folio_mapcount(folio), folio); local
2017 pages_per_huge_page(h), folio); local
2019 pages_per_huge_page(h), folio); local
2050 init_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument
2059 __prep_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument
2065 prep_new_hugetlb_folio(struct hstate *h, struct folio *folio, int nid) argument
2073 __prep_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument
2150 prep_compound_gigantic_folio(struct folio *folio, unsigned int order) argument
2156 prep_compound_gigantic_folio_for_demote(struct folio *folio, unsigned int order) argument
2249 struct folio *folio; local
2283 struct folio *folio; local
2303 struct folio *folio; local
2318 struct folio *folio, *tmp_f; local
2345 struct folio *folio; local
2367 struct folio *folio = NULL; local
2404 struct folio *folio = page_folio(page); local
2515 struct folio *folio = NULL; local
2556 struct folio *folio; local
2583 struct folio *folio = NULL; local
2612 struct folio *folio; local
2634 struct folio *folio, *tmp; variable in typeref:struct:folio
2750 struct folio *folio; local
2932 restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma, unsigned long address, struct folio *folio) argument
3099 struct folio *folio = page_folio(page); local
3137 struct folio *folio; local
3258 pages_per_huge_page(h), folio); local
3335 hugetlb_folio_init_tail_vmemmap(struct folio *folio, unsigned long start_page_number, unsigned long end_page_number) argument
3355 hugetlb_folio_init_vmemmap(struct folio *folio, struct hstate *h, unsigned long nr_pages) argument
3375 struct folio *folio, *tmp_f; local
3412 struct folio *folio = (void *)page; local
3478 struct folio *folio; local
3538 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY], local
3706 struct folio *folio, *next; local
3769 struct folio *folio; local
3936 demote_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument
4013 struct folio *folio; variable in typeref:struct:folio
6124 struct folio *folio; local
6133 hugetlb_add_to_page_cache(struct folio *folio, struct address_space *mapping, pgoff_t idx) argument
6207 struct folio *folio; local
6439 struct folio *folio = NULL; local
6640 struct folio *folio; local
6674 struct folio *folio; local
7592 isolate_hugetlb(struct folio *folio, struct list_head *list) argument
7610 get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison) argument
7640 folio_putback_active_hugetlb(struct folio *folio) argument
[all...]
/linux-master/arch/openrisc/mm/
H A Dcache.c46 struct folio *folio = page_folio(pfn_to_page(pfn)); local
47 int dirty = !test_and_set_bit(PG_dc_clean, &folio->flags);
55 unsigned int nr = folio_nr_pages(folio);
58 sync_icache_dcache(folio_page(folio, nr));
/linux-master/include/linux/
H A Dhighmem.h99 * kmap_local_folio - Map a page in this folio for temporary usage
100 * @folio: The folio containing the page.
101 * @offset: The byte offset within the folio which identifies the page.
132 static inline void *kmap_local_folio(struct folio *folio, size_t offset);
220 * Return: A folio containing one allocated and zeroed page or NULL if
224 struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
227 struct folio *folio; local
449 memcpy_from_folio(char *to, struct folio *folio, size_t offset, size_t len) argument
477 memcpy_to_folio(struct folio *folio, size_t offset, const char *from, size_t len) argument
514 folio_zero_tail(struct folio *folio, size_t offset, void *kaddr) argument
550 folio_fill_tail(struct folio *folio, size_t offset, const char *from, size_t len) argument
588 memcpy_from_file_folio(char *to, struct folio *folio, loff_t pos, size_t len) argument
614 folio_zero_segments(struct folio *folio, size_t start1, size_t xend1, size_t start2, size_t xend2) argument
626 folio_zero_segment(struct folio *folio, size_t start, size_t xend) argument
638 folio_zero_range(struct folio *folio, size_t start, size_t length) argument
653 folio_release_kmap(struct folio *folio, void *addr) argument
[all...]
H A Dmemcontrol.h362 static inline bool folio_memcg_kmem(struct folio *folio);
377 * __folio_memcg - Get the memory cgroup associated with a non-kmem folio
378 * @folio: Pointer to the folio.
380 * Returns a pointer to the memory cgroup associated with the folio,
381 * or NULL. This function assumes that the folio is known to have a
386 static inline struct mem_cgroup *__folio_memcg(struct folio *folio) argument
388 unsigned long memcg_data = folio
390 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); local
407 __folio_objcg(struct folio *folio) argument
411 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); local
439 folio_memcg(struct folio *folio) argument
462 folio_memcg_rcu(struct folio *folio) argument
466 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); local
501 folio_memcg_check(struct folio *folio) argument
552 folio_memcg_kmem(struct folio *folio) argument
561 folio_memcg_kmem(struct folio *folio) argument
686 mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp) argument
709 mem_cgroup_uncharge(struct folio *folio) argument
770 folio_lruvec(struct folio *folio) argument
793 lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) argument
1100 count_memcg_folio_events(struct folio *folio, enum vm_event_item idx, unsigned long nr) argument
1174 folio_memcg(struct folio *folio) argument
1184 folio_memcg_rcu(struct folio *folio) argument
1190 folio_memcg_check(struct folio *folio) argument
1205 folio_memcg_kmem(struct folio *folio) argument
1265 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
1270 mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp) argument
1282 mem_cgroup_swapin_charge_folio(struct folio *folio, struct mm_struct *mm, gfp_t gfp, swp_entry_t entry) argument
1292 mem_cgroup_uncharge(struct folio *folio) argument
1320 folio_lruvec(struct folio *folio) argument
1327 lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) argument
1376 folio_lruvec_lock(struct folio *folio) argument
1384 folio_lruvec_lock_irq(struct folio *folio) argument
1392 folio_lruvec_lock_irqsave(struct folio *folio, unsigned long *flagsp) argument
1485 folio_memcg_lock(struct folio *folio) argument
1489 folio_memcg_unlock(struct folio *folio) argument
1612 count_memcg_folio_events(struct folio *folio, enum vm_event_item idx, unsigned long nr) argument
1675 folio_matches_lruvec(struct folio *folio, struct lruvec *lruvec) argument
1683 folio_lruvec_relock_irq(struct folio *folio, struct lruvec *locked_lruvec) argument
1697 folio_lruvec_relock_irqsave(struct folio *folio, struct lruvec **lruvecp, unsigned long *flags) argument
1720 mem_cgroup_track_foreign_dirty(struct folio *folio, struct bdi_writeback *wb) argument
1750 mem_cgroup_track_foreign_dirty(struct folio *folio, struct bdi_writeback *wb) argument
1908 get_obj_cgroup_from_folio(struct folio *folio) argument
[all...]
H A Dvmstat.h418 static inline void __zone_stat_mod_folio(struct folio *folio, argument
421 __mod_zone_page_state(folio_zone(folio), item, nr);
424 static inline void __zone_stat_add_folio(struct folio *folio, argument
427 __mod_zone_page_state(folio_zone(folio), item, folio_nr_pages(folio));
430 static inline void __zone_stat_sub_folio(struct folio *folio, argument
433 __mod_zone_page_state(folio_zone(folio), ite
436 zone_stat_mod_folio(struct folio *folio, enum zone_stat_item item, long nr) argument
442 zone_stat_add_folio(struct folio *folio, enum zone_stat_item item) argument
448 zone_stat_sub_folio(struct folio *folio, enum zone_stat_item item) argument
454 __node_stat_mod_folio(struct folio *folio, enum node_stat_item item, long nr) argument
460 __node_stat_add_folio(struct folio *folio, enum node_stat_item item) argument
466 __node_stat_sub_folio(struct folio *folio, enum node_stat_item item) argument
472 node_stat_mod_folio(struct folio *folio, enum node_stat_item item, long nr) argument
478 node_stat_add_folio(struct folio *folio, enum node_stat_item item) argument
484 node_stat_sub_folio(struct folio *folio, enum node_stat_item item) argument
562 lruvec_stat_mod_folio(struct folio *folio, enum node_stat_item idx, int val) argument
592 __lruvec_stat_mod_folio(struct folio *folio, enum node_stat_item idx, int val) argument
598 lruvec_stat_mod_folio(struct folio *folio, enum node_stat_item idx, int val) argument
612 __lruvec_stat_add_folio(struct folio *folio, enum node_stat_item idx) argument
618 __lruvec_stat_sub_folio(struct folio *folio, enum node_stat_item idx) argument
624 lruvec_stat_add_folio(struct folio *folio, enum node_stat_item idx) argument
630 lruvec_stat_sub_folio(struct folio *folio, enum node_stat_item idx) argument
[all...]
H A Dnetfs.h28 #define folio_test_fscache(folio) folio_test_private_2(folio)
36 * folio_start_fscache - Start an fscache write on a folio.
37 * @folio: The folio.
39 * Call this function before writing a folio to a local cache. Starting a
42 static inline void folio_start_fscache(struct folio *folio) argument
44 VM_BUG_ON_FOLIO(folio_test_private_2(folio), folio); local
56 folio_end_fscache(struct folio *folio) argument
69 folio_wait_fscache(struct folio *folio) argument
87 folio_wait_fscache_killable(struct folio *folio) argument
169 netfs_folio_info(struct folio *folio) argument
178 netfs_folio_group(struct folio *folio) argument
[all...]
H A Dpage-flags.h144 * Depending on the way an anonymous folio can be mapped into a page
147 * tail pages of an anonymous folio. For now, we only expect it to be
191 /* At least one page in this folio has the hwpoison flag set */
256 * page_folio - Converts from page to folio.
259 * Every page is part of a folio. This function cannot be called on a
263 * does not hold a reference, this call may race with a folio split, so
264 * it should re-check the folio still contains this page after gaining
265 * a reference on the folio.
266 * Return: The folio which contains this page.
269 const struct page *: (const struct folio *)_compound_hea
308 const_folio_flags(const struct folio *folio, unsigned n) argument
318 folio_flags(struct folio *folio, unsigned n) argument
581 folio_test_swapcache(const struct folio *folio) argument
686 folio_mapping_flags(const struct folio *folio) argument
696 folio_test_anon(const struct folio *folio) argument
706 __folio_test_movable(const struct folio *folio) argument
725 folio_test_ksm(const struct folio *folio) argument
754 folio_xor_flags_has_waiters(struct folio *folio, unsigned long mask) argument
770 folio_test_uptodate(const struct folio *folio) argument
792 __folio_mark_uptodate(struct folio *folio) argument
798 folio_mark_uptodate(struct folio *folio) argument
829 folio_test_head(const struct folio *folio) argument
850 folio_test_large(const struct folio *folio) argument
1157 folio_has_private(const struct folio *folio) argument
[all...]
/linux-master/fs/affs/
H A Dsymlink.c14 static int affs_symlink_read_folio(struct file *file, struct folio *folio) argument
17 struct inode *inode = folio->mapping->host;
18 char *link = folio_address(folio);
60 folio_mark_uptodate(folio);
61 folio_unlock(folio);
64 folio_unlock(folio);
/linux-master/fs/nilfs2/
H A Dfile.c48 struct folio *folio = page_folio(vmf->page); local
58 folio_lock(folio);
59 if (folio->mapping != inode->i_mapping ||
60 folio_pos(folio) >= i_size_read(inode) ||
61 !folio_test_uptodate(folio)) {
62 folio_unlock(folio);
68 * check to see if the folio is mapped already (no holes)
70 if (folio_test_mappedtodisk(folio))
73 head = folio_buffers(folio);
[all...]
/linux-master/fs/xfs/scrub/
H A Dxfile.h24 #define XFILE_ALLOC (1 << 0) /* allocate folio if not present */
25 struct folio *xfile_get_folio(struct xfile *xf, loff_t offset, size_t len,
27 void xfile_put_folio(struct xfile *xf, struct folio *folio);
/linux-master/arch/nios2/mm/
H A Dcacheflush.c74 static void flush_aliases(struct address_space *mapping, struct folio *folio) argument
80 unsigned long nr = folio_nr_pages(folio);
82 pgoff = folio->index;
163 static void __flush_dcache_folio(struct folio *folio) argument
170 unsigned long start = (unsigned long)folio_address(folio);
172 __flush_dcache(start, start + folio_size(folio));
175 void flush_dcache_folio(struct folio *folio) argument
214 struct folio *folio; local
[all...]
/linux-master/fs/bcachefs/
H A Dfs-io-pagecache.c19 struct folio *f;
126 struct bch_folio *__bch2_folio_create(struct folio *folio, gfp_t gfp) argument
132 folio_sectors(folio), gfp);
137 folio_attach_private(folio, s);
141 struct bch_folio *bch2_folio_create(struct folio *folio, gfp_t gfp) argument
143 return bch2_folio(folio) ?: __bch2_folio_create(folio, gfp);
155 static void __bch2_folio_set(struct folio *foli argument
222 struct folio *folio = fs[folio_idx]; local
286 struct folio *folio = fbatch.folios[i]; local
331 struct folio *folio = fbatch.folios[i]; local
380 bch2_get_folio_disk_reservation(struct bch_fs *c, struct bch_inode_info *inode, struct folio *folio, bool check_enospc) argument
422 bch2_folio_reservation_get(struct bch_fs *c, struct bch_inode_info *inode, struct folio *folio, struct bch2_folio_reservation *res, unsigned offset, unsigned len) argument
468 bch2_clear_folio_bits(struct folio *folio) argument
497 bch2_set_folio_dirty(struct bch_fs *c, struct bch_inode_info *inode, struct folio *folio, struct bch2_folio_reservation *res, unsigned offset, unsigned len) argument
581 struct folio *folio = page_folio(vmf->page); local
634 bch2_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument
642 bch2_release_folio(struct folio *folio, gfp_t gfp_mask) argument
653 folio_data_offset(struct folio *folio, loff_t pos, unsigned min_replicas) argument
687 struct folio *folio = fbatch.folios[i]; local
726 struct folio *folio; local
[all...]
H A Dfs-io-buffered.h7 int bch2_read_single_folio(struct folio *, struct address_space *);
8 int bch2_read_folio(struct file *, struct folio *);
/linux-master/fs/
H A Dmpage.c53 folio_set_error(fi.folio);
55 folio_mark_uptodate(fi.folio);
56 folio_unlock(fi.folio);
69 folio_set_error(fi.folio);
70 mapping_set_error(fi.folio->mapping, err);
72 folio_end_writeback(fi.folio);
104 static void map_buffer_to_folio(struct folio *folio, struct buffer_head *bh, argument
107 struct inode *inode = folio->mapping->host;
111 head = folio_buffers(folio);
140 struct folio *folio; member in struct:mpage_readpage_args
160 struct folio *folio = args->folio; local
181 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); local
371 struct folio *folio; local
391 mpage_read_folio(struct folio *folio, get_block_t get_block) argument
433 clean_buffers(struct folio *folio, unsigned first_unmapped) argument
458 __mpage_writepage(struct folio *folio, struct writeback_control *wbc, void *data) argument
[all...]
/linux-master/include/trace/events/
H A Dfilemap.h18 TP_PROTO(struct folio *folio),
20 TP_ARGS(folio),
31 __entry->pfn = folio_pfn(folio);
32 __entry->i_ino = folio->mapping->host->i_ino;
33 __entry->index = folio->index;
34 if (folio->mapping->host->i_sb)
35 __entry->s_dev = folio->mapping->host->i_sb->s_dev;
37 __entry->s_dev = folio->mapping->host->i_rdev;
38 __entry->order = folio_order(folio);
[all...]
/linux-master/include/linux/sched/
H A Dnuma_balancing.h33 bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio,
51 struct folio *folio, int src_nid, int dst_cpu)
50 should_numa_migrate_memory(struct task_struct *p, struct folio *folio, int src_nid, int dst_cpu) argument
/linux-master/arch/riscv/include/asm/
H A Dcacheflush.h18 static inline void flush_dcache_folio(struct folio *folio) argument
20 if (test_bit(PG_dcache_clean, &folio->flags))
21 clear_bit(PG_dcache_clean, &folio->flags);
/linux-master/arch/mips/include/asm/
H A Dcacheflush.h39 #define folio_test_dcache_dirty(folio) \
40 test_bit(PG_dcache_dirty, &(folio)->flags)
41 #define folio_set_dcache_dirty(folio) \
42 set_bit(PG_dcache_dirty, &(folio)->flags)
43 #define folio_clear_dcache_dirty(folio) \
44 clear_bit(PG_dcache_dirty, &(folio)->flags)
56 static inline void flush_dcache_folio(struct folio *folio) argument
59 __flush_dcache_pages(&folio->page, folio_nr_pages(folio));
[all...]
/linux-master/fs/afs/
H A Ddir_edit.c108 * Get a new directory folio.
110 static struct folio *afs_dir_get_folio(struct afs_vnode *vnode, pgoff_t index)
113 struct folio *folio; local
115 folio = __filemap_get_folio(mapping, index,
118 if (IS_ERR(folio)) {
122 if (!folio_test_private(folio))
123 folio_attach_private(folio, (void *)1);
124 return folio;
212 struct folio *folio local
379 struct folio *folio0, *folio; local
[all...]
/linux-master/fs/nfs/
H A Dfscache.h87 extern int nfs_netfs_folio_unlock(struct folio *folio);
100 extern int nfs_netfs_read_folio(struct file *file, struct folio *folio);
102 static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) argument
104 if (folio_test_fscache(folio)) {
107 folio_wait_fscache(folio);
109 fscache_note_page_release(netfs_i_cookie(netfs_inode(folio->mapping->host)));
166 static inline int nfs_netfs_folio_unlock(struct folio *foli argument
181 nfs_netfs_read_folio(struct file *file, struct folio *folio) argument
186 nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) argument
[all...]

Completed in 374 milliseconds

1234567891011>>