Searched refs:folio (Results 1 - 25 of 388) sorted by relevance

1234567891011>>

/linux-master/tools/virtio/linux/
H A Dmm_types.h1 struct folio { struct
/linux-master/mm/
H A Dswap.c60 * The following folio batches are grouped together because they are protected
77 static void __page_cache_release(struct folio *folio, struct lruvec **lruvecp, argument
80 if (folio_test_lru(folio)) {
81 folio_lruvec_relock_irqsave(folio, lruvecp, flagsp);
82 lruvec_del_folio(*lruvecp, folio);
83 __folio_clear_lru_flags(folio);
92 if (unlikely(folio_test_mlocked(folio))) {
93 long nr_pages = folio_nr_pages(folio);
95 __folio_clear_mlocked(folio);
105 page_cache_release(struct folio *folio) argument
115 __folio_put(struct folio *folio) argument
142 struct folio *folio, *next; local
166 lru_add_fn(struct lruvec *lruvec, struct folio *folio) argument
171 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local
213 struct folio *folio = fbatch->folios[i]; local
230 folio_batch_add_and_move(struct folio_batch *fbatch, struct folio *folio, move_fn_t move_fn) argument
239 lru_move_tail_fn(struct lruvec *lruvec, struct folio *folio) argument
256 folio_rotate_reclaimable(struct folio *folio) argument
323 lru_note_cost_refault(struct folio *folio) argument
329 folio_activate_fn(struct lruvec *lruvec, struct folio *folio) argument
354 folio_activate(struct folio *folio) argument
373 folio_activate(struct folio *folio) argument
386 __lru_cache_activate_folio(struct folio *folio) argument
417 folio_inc_refs(struct folio *folio) argument
445 folio_inc_refs(struct folio *folio) argument
463 folio_mark_accessed(struct folio *folio) argument
506 folio_add_lru(struct folio *folio) argument
512 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local
535 folio_add_lru_vma(struct folio *folio, struct vm_area_struct *vma) argument
537 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local
566 lru_deactivate_file_fn(struct lruvec *lruvec, struct folio *folio) argument
607 lru_deactivate_fn(struct lruvec *lruvec, struct folio *folio) argument
623 lru_lazyfree_fn(struct lruvec *lruvec, struct folio *folio) argument
695 deactivate_file_folio(struct folio *folio) argument
718 folio_deactivate(struct folio *folio) argument
739 folio_mark_lazyfree(struct folio *folio) argument
975 struct folio *folio = folios->folios[i]; local
1050 struct folio *folio = page_folio(encoded_page_ptr(encoded[i])); local
1102 struct folio *folio = fbatch->folios[i]; local
[all...]
H A Dpage_idle.c34 static struct folio *page_idle_get_folio(unsigned long pfn)
37 struct folio *folio; local
42 folio = page_folio(page);
43 if (!folio_test_lru(folio) || !folio_try_get(folio))
45 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) {
46 folio_put(folio);
47 folio
52 page_idle_clear_pte_refs_one(struct folio *folio, struct vm_area_struct *vma, unsigned long addr, void *arg) argument
89 page_idle_clear_pte_refs(struct folio *folio) argument
119 struct folio *folio; local
164 struct folio *folio; local
[all...]
H A Dswap.h13 void swap_read_folio(struct folio *folio, bool do_poll,
23 void __swap_writepage(struct folio *folio, struct writeback_control *wbc);
35 bool add_to_swap(struct folio *folio);
37 int add_to_swap_cache(struct folio *folio, swp_entry_t entry,
39 void __delete_from_swap_cache(struct folio *folio,
61 folio_swap_flags(struct folio *folio) argument
67 swap_read_folio(struct folio *folio, bool do_poll, struct swap_iocb **plug) argument
118 add_to_swap(struct folio *folio) argument
128 add_to_swap_cache(struct folio *folio, swp_entry_t entry, gfp_t gfp_mask, void **shadowp) argument
134 __delete_from_swap_cache(struct folio *folio, swp_entry_t entry, void *shadow) argument
139 delete_from_swap_cache(struct folio *folio) argument
148 folio_swap_flags(struct folio *folio) argument
[all...]
H A Dtruncate.c83 struct folio *folio = fbatch->folios[i]; local
86 if (!xa_is_value(folio)) {
87 fbatch->folios[j++] = folio;
96 __clear_shadow_entry(mapping, index, folio);
139 * folio_invalidate - Invalidate part or all of a folio.
140 * @folio: The folio which is affected.
144 * folio_invalidate() is called when all or part of the folio has become
153 void folio_invalidate(struct folio *foli argument
172 truncate_cleanup_folio(struct folio *folio) argument
189 truncate_inode_folio(struct address_space *mapping, struct folio *folio) argument
210 truncate_inode_partial_folio(struct folio *folio, loff_t start, loff_t end) argument
253 generic_error_remove_folio(struct address_space *mapping, struct folio *folio) argument
279 mapping_evict_folio(struct address_space *mapping, struct folio *folio) argument
329 struct folio *folio; local
405 struct folio *folio = fbatch.folios[i]; local
501 struct folio *folio = fbatch.folios[i]; local
560 invalidate_complete_folio2(struct address_space *mapping, struct folio *folio) argument
589 folio_launder(struct address_space *mapping, struct folio *folio) argument
627 struct folio *folio = fbatch.folios[i]; local
784 struct folio *folio; local
[all...]
/linux-master/include/trace/events/
H A Dpagemap.h19 #define trace_pagemap_flags(folio) ( \
20 (folio_test_anon(folio) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \
21 (folio_mapped(folio) ? PAGEMAP_MAPPED : 0) | \
22 (folio_test_swapcache(folio) ? PAGEMAP_SWAPCACHE : 0) | \
23 (folio_test_swapbacked(folio) ? PAGEMAP_SWAPBACKED : 0) | \
24 (folio_test_mappedtodisk(folio) ? PAGEMAP_MAPPEDDISK : 0) | \
25 (folio_test_private(folio) ? PAGEMAP_BUFFERS : 0) \
30 TP_PROTO(struct folio *folio),
32 TP_ARGS(folio),
[all...]
/linux-master/include/linux/
H A Dhugetlb_cgroup.h61 __hugetlb_cgroup_from_folio(struct folio *folio, bool rsvd) argument
63 VM_BUG_ON_FOLIO(!folio_test_hugetlb(folio), folio);
65 return folio->_hugetlb_cgroup_rsvd;
67 return folio->_hugetlb_cgroup;
70 static inline struct hugetlb_cgroup *hugetlb_cgroup_from_folio(struct folio *folio) argument
72 return __hugetlb_cgroup_from_folio(folio, false);
76 hugetlb_cgroup_from_folio_rsvd(struct folio *foli argument
81 __set_hugetlb_cgroup(struct folio *folio, struct hugetlb_cgroup *h_cg, bool rsvd) argument
91 set_hugetlb_cgroup(struct folio *folio, struct hugetlb_cgroup *h_cg) argument
97 set_hugetlb_cgroup_rsvd(struct folio *folio, struct hugetlb_cgroup *h_cg) argument
167 hugetlb_cgroup_from_folio(struct folio *folio) argument
173 hugetlb_cgroup_from_folio_rsvd(struct folio *folio) argument
178 set_hugetlb_cgroup(struct folio *folio, struct hugetlb_cgroup *h_cg) argument
183 set_hugetlb_cgroup_rsvd(struct folio *folio, struct hugetlb_cgroup *h_cg) argument
220 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument
227 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument
233 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) argument
238 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) argument
[all...]
H A Dcacheflush.h7 struct folio;
11 void flush_dcache_folio(struct folio *folio);
14 static inline void flush_dcache_folio(struct folio *folio) argument
H A Drmap.h174 struct anon_vma *folio_get_anon_vma(struct folio *folio);
197 static inline void __folio_rmap_sanity_checks(struct folio *folio, argument
201 VM_WARN_ON_FOLIO(folio_test_hugetlb(folio), folio); local
213 VM_WARN_ON_FOLIO(page_folio(page) != folio, folio); local
214 VM_WARN_ON_FOLIO(page_folio(page + nr_pages - 1) != folio, folio);
225 VM_WARN_ON_FOLIO(folio_nr_pages(folio) != HPAGE_PMD_NR, folio); local
264 hugetlb_try_dup_anon_rmap(struct folio *folio, struct vm_area_struct *vma) argument
281 hugetlb_try_share_anon_rmap(struct folio *folio) argument
304 hugetlb_add_file_rmap(struct folio *folio) argument
307 VM_WARN_ON_FOLIO(folio_test_anon(folio), folio); local
313 hugetlb_remove_rmap(struct folio *folio) argument
321 __folio_dup_file_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
357 folio_dup_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages) argument
363 folio_dup_file_rmap_pte(struct folio *folio, struct page *page) argument
378 folio_dup_file_rmap_pmd(struct folio *folio, struct page *page) argument
388 __folio_try_dup_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma, enum rmap_level level) argument
472 folio_try_dup_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma) argument
479 folio_try_dup_anon_rmap_pte(struct folio *folio, struct page *page, struct vm_area_struct *src_vma) argument
508 folio_try_dup_anon_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *src_vma) argument
520 __folio_try_share_anon_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
613 folio_try_share_anon_rmap_pte(struct folio *folio, struct page *page) argument
642 folio_try_share_anon_rmap_pmd(struct folio *folio, struct page *page) argument
772 folio_referenced(struct folio *folio, int is_locked, struct mem_cgroup *memcg, unsigned long *vm_flags) argument
780 try_to_unmap(struct folio *folio, enum ttu_flags flags) argument
784 folio_mkclean(struct folio *folio) argument
[all...]
H A Dzswap.h30 bool zswap_store(struct folio *folio);
31 bool zswap_load(struct folio *folio);
37 void zswap_folio_swapin(struct folio *folio);
43 static inline bool zswap_store(struct folio *folio) argument
48 static inline bool zswap_load(struct folio *folio) argument
61 zswap_folio_swapin(struct folio *folio) argument
[all...]
H A Dpage_idle.h14 static inline bool folio_test_young(const struct folio *folio) argument
16 struct page_ext *page_ext = page_ext_get(&folio->page);
28 static inline void folio_set_young(struct folio *folio) argument
30 struct page_ext *page_ext = page_ext_get(&folio->page);
39 static inline bool folio_test_clear_young(struct folio *folio) argument
41 struct page_ext *page_ext = page_ext_get(&folio->page);
53 static inline bool folio_test_idle(const struct folio *foli argument
67 folio_set_idle(struct folio *folio) argument
78 folio_clear_idle(struct folio *folio) argument
[all...]
H A Dmigrate.h10 typedef struct folio *new_folio_t(struct folio *folio, unsigned long private);
11 typedef void free_folio_t(struct folio *folio, unsigned long private);
66 int migrate_folio_extra(struct address_space *mapping, struct folio *dst,
67 struct folio *src, enum migrate_mode mode, int extra_count);
68 int migrate_folio(struct address_space *mapping, struct folio *dst,
69 struct folio *src, enum migrate_mode mode);
73 struct folio *alloc_migration_targe
121 folio_test_movable(struct folio *folio) argument
127 folio_movable_ops(struct folio *folio) argument
148 migrate_misplaced_folio(struct folio *folio, struct vm_area_struct *vma, int node) argument
[all...]
H A Dpage_ref.h71 * folio_ref_count - The reference count on this folio.
72 * @folio: The folio.
76 * folio refcount:
85 * Return: The number of references to this folio.
87 static inline int folio_ref_count(const struct folio *folio) argument
89 return page_ref_count(&folio->page);
104 static inline void folio_set_count(struct folio *folio, in argument
125 folio_ref_add(struct folio *folio, int nr) argument
137 folio_ref_sub(struct folio *folio, int nr) argument
142 folio_ref_sub_return(struct folio *folio, int nr) argument
158 folio_ref_inc(struct folio *folio) argument
170 folio_ref_dec(struct folio *folio) argument
184 folio_ref_sub_and_test(struct folio *folio, int nr) argument
198 folio_ref_inc_return(struct folio *folio) argument
212 folio_ref_dec_and_test(struct folio *folio) argument
226 folio_ref_dec_return(struct folio *folio) argument
240 folio_ref_add_unless(struct folio *folio, int nr, int u) argument
256 folio_try_get(struct folio *folio) argument
261 folio_ref_try_add_rcu(struct folio *folio, int count) argument
272 VM_BUG_ON_FOLIO(folio_ref_count(folio) == 0, folio); local
306 folio_try_get_rcu(struct folio *folio) argument
320 folio_ref_freeze(struct folio *folio, int count) argument
335 folio_ref_unfreeze(struct folio *folio, int count) argument
[all...]
H A Dmpage.h19 int mpage_read_folio(struct folio *folio, get_block_t get_block);
H A Dpagemap.h210 folio contents */
366 * Large folio support currently depends on THP. These dependencies are
405 struct address_space *folio_mapping(struct folio *);
406 struct address_space *swapcache_mapping(struct folio *);
409 * folio_file_mapping - Find the mapping this folio belongs to.
410 * @folio: The folio.
420 static inline struct address_space *folio_file_mapping(struct folio *folio) argument
422 if (unlikely(folio_test_swapcache(folio)))
440 folio_flush_mapping(struct folio *folio) argument
462 folio_inode(struct folio *folio) argument
475 folio_attach_private(struct folio *folio, void *data) argument
493 folio_change_private(struct folio *folio, void *data) argument
510 folio_detach_private(struct folio *folio) argument
796 folio_index(struct folio *folio) argument
809 folio_next_index(struct folio *folio) argument
824 folio_file_page(struct folio *folio, pgoff_t index) argument
839 folio_contains(struct folio *folio, pgoff_t index) argument
932 folio_pos(struct folio *folio) argument
944 folio_file_pos(struct folio *folio) argument
952 folio_pgoff(struct folio *folio) argument
967 struct folio *folio; member in struct:wait_page_key
973 struct folio *folio; member in struct:wait_page_queue
1009 folio_trylock(struct folio *folio) argument
1044 folio_lock(struct folio *folio) argument
1064 struct folio *folio; local
1082 folio_lock_killable(struct folio *folio) argument
1097 folio_lock_or_retry(struct folio *folio, struct vm_fault *vmf) argument
1120 folio_wait_locked(struct folio *folio) argument
1126 folio_wait_locked_killable(struct folio *folio) argument
1149 folio_cancel_dirty(struct folio *folio) argument
1312 page_cache_async_readahead(struct address_space *mapping, struct file_ra_state *ra, struct file *file, struct folio *folio, pgoff_t index, unsigned long req_count) argument
1322 struct folio *folio; local
1351 struct folio *folio = __readahead_folio(ractl); local
1366 struct folio *folio = __readahead_folio(ractl); local
1475 folio_mkwrite_check_truncate(struct folio *folio, struct inode *inode) argument
1535 i_blocks_per_folio(struct inode *inode, struct folio *folio) argument
[all...]
/linux-master/fs/gfs2/
H A Daops.h12 void gfs2_trans_add_databufs(struct gfs2_inode *ip, struct folio *folio,
/linux-master/fs/freevxfs/
H A Dvxfs_immed.c19 * @folio: folio to fill in.
23 * file that hosts @folio into the pagecache.
29 * @folio is locked and will be unlocked.
31 static int vxfs_immed_read_folio(struct file *fp, struct folio *folio) argument
33 struct vxfs_inode_info *vip = VXFS_INO(folio->mapping->host);
34 void *src = vip->vii_immed.vi_immed + folio_pos(folio);
37 for (i = 0; i < folio_nr_pages(folio); i++) {
38 memcpy_to_page(folio_page(folio,
[all...]
/linux-master/fs/bcachefs/
H A Dfs-io-pagecache.h7 typedef DARRAY(struct folio *) folios;
14 * Use u64 for the end pos and sector helpers because if the folio covers the
15 * max supported range of the mapping, the start offset of the next folio
19 static inline u64 folio_end_pos(struct folio *folio) argument
21 return folio_pos(folio) + folio_size(folio);
24 static inline size_t folio_sectors(struct folio *folio) argument
26 return PAGE_SECTORS << folio_order(folio);
29 folio_sector(struct folio *folio) argument
34 folio_end_sector(struct folio *folio) argument
72 bch2_folio_sector_set(struct folio *folio, struct bch_folio *s, unsigned i, unsigned n) argument
80 folio_pos_to_s(struct folio *folio, loff_t pos) argument
89 __bch2_folio_release(struct folio *folio) argument
94 bch2_folio_release(struct folio *folio) argument
100 __bch2_folio(struct folio *folio) argument
107 bch2_folio(struct folio *folio) argument
[all...]
/linux-master/fs/btrfs/
H A Dsubpage.h10 struct folio;
46 * For locked bitmaps, normally it's subpage representation for folio
49 * - Metadata doesn't really lock the folio
95 struct folio *folio, enum btrfs_subpage_type type);
96 void btrfs_detach_subpage(const struct btrfs_fs_info *fs_info, struct folio *folio);
103 void btrfs_folio_inc_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *folio);
104 void btrfs_folio_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *foli
[all...]
H A Dsubpage.c121 struct folio *folio, enum btrfs_subpage_type type)
129 if (folio->mapping)
130 ASSERT(folio_test_locked(folio));
132 /* Either not subpage, or the folio already has private attached. */
133 if (!btrfs_is_subpage(fs_info, folio->mapping) || folio_test_private(folio))
140 folio_attach_private(folio, subpage);
144 void btrfs_detach_subpage(const struct btrfs_fs_info *fs_info, struct folio *folio) argument
120 btrfs_attach_subpage(const struct btrfs_fs_info *fs_info, struct folio *folio, enum btrfs_subpage_type type) argument
195 btrfs_folio_inc_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *folio) argument
209 btrfs_folio_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *folio) argument
224 btrfs_subpage_assert(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
253 btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
275 btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
309 btrfs_subpage_clamp_range(struct folio *folio, u64 *start, u32 *len) argument
327 btrfs_subpage_start_writer(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
347 btrfs_subpage_end_and_test_writer(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
390 btrfs_folio_start_writer_lock(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
407 btrfs_folio_end_writer_lock(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
429 btrfs_subpage_set_uptodate(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
444 btrfs_subpage_clear_uptodate(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
458 btrfs_subpage_set_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
482 btrfs_subpage_clear_and_test_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
499 btrfs_subpage_clear_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
509 btrfs_subpage_set_writeback(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
524 btrfs_subpage_clear_writeback(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
541 btrfs_subpage_set_ordered(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
555 btrfs_subpage_clear_ordered(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
570 btrfs_subpage_set_checked(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
585 btrfs_subpage_clear_checked(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
706 btrfs_folio_assert_not_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio) argument
736 btrfs_folio_unlock_writer(struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
772 btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument
[all...]
/linux-master/fs/netfs/
H A Dmisc.c12 * Attach a folio to the buffer and maybe set marks on it to say that we need
13 * to put the folio later and twiddle the pagecache flags.
16 struct folio *folio, unsigned int flags,
19 XA_STATE_ORDER(xas, xa, index, folio_order(folio));
24 xas_store(&xas, folio);
50 struct folio *folio; local
57 /* TODO: Figure out what order folio can be allocated here */
58 folio
15 netfs_xa_store_and_mark(struct xarray *xa, unsigned long index, struct folio *folio, unsigned int flags, gfp_t gfp_mask) argument
81 struct folio *folio; local
103 netfs_dirty_folio(struct address_space *mapping, struct folio *folio) argument
178 netfs_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument
238 netfs_release_folio(struct folio *folio, gfp_t gfp) argument
[all...]
/linux-master/mm/damon/
H A Dops-common.c22 struct folio *damon_get_folio(unsigned long pfn)
25 struct folio *folio; local
30 folio = page_folio(page);
31 if (!folio_test_lru(folio) || !folio_try_get(folio))
33 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) {
34 folio_put(folio);
35 folio
42 struct folio *folio = damon_get_folio(pte_pfn(ptep_get(pte))); local
57 struct folio *folio = damon_get_folio(pmd_pfn(pmdp_get(pmd))); local
[all...]
H A Dpaddr.c19 static bool damon_folio_mkold_one(struct folio *folio, argument
22 DEFINE_FOLIO_VMA_WALK(pvmw, folio, vma, addr, 0);
34 static void damon_folio_mkold(struct folio *folio) argument
42 if (!folio_mapped(folio) || !folio_raw_mapping(folio)) {
43 folio_set_idle(folio);
47 need_lock = !folio_test_anon(folio) || folio_test_ksm(folio);
60 struct folio *folio = damon_get_folio(PHYS_PFN(paddr)); local
87 damon_folio_young_one(struct folio *folio, struct vm_area_struct *vma, unsigned long addr, void *arg) argument
119 damon_folio_young(struct folio *folio) argument
150 struct folio *folio = damon_get_folio(PHYS_PFN(paddr)); local
198 __damos_pa_filter_out(struct damos_filter *filter, struct folio *folio) argument
232 damos_pa_filter_out(struct damos *scheme, struct folio *folio) argument
265 struct folio *folio = damon_get_folio(PHYS_PFN(addr)); local
297 struct folio *folio = damon_get_folio(PHYS_PFN(addr)); local
[all...]
/linux-master/fs/coda/
H A Dsymlink.c23 static int coda_symlink_filler(struct file *file, struct folio *folio) argument
25 struct inode *inode = folio->mapping->host;
29 char *p = folio_address(folio);
36 folio_mark_uptodate(folio);
37 folio_unlock(folio);
41 folio_set_error(folio);
42 folio_unlock(folio);
/linux-master/fs/nilfs2/
H A Dpage.c28 static struct buffer_head *__nilfs_get_folio_block(struct folio *folio, argument
34 struct buffer_head *bh = folio_buffers(folio);
37 bh = create_empty_buffers(folio, 1 << blkbits, b_state);
54 struct folio *folio; local
57 folio = filemap_grab_folio(mapping, index);
58 if (IS_ERR(folio))
61 bh = __nilfs_get_folio_block(folio, blkoff, index, blkbits, b_state);
63 folio_unlock(folio);
76 struct folio *folio = bh->b_folio; local
140 nilfs_folio_buffers_clean(struct folio *folio) argument
153 nilfs_folio_bug(struct folio *folio) argument
258 struct folio *folio = fbatch.folios[i], *dfolio; local
312 struct folio *folio = fbatch.folios[i], *dfolio; local
373 struct folio *folio = fbatch.folios[i]; local
397 nilfs_clear_folio_dirty(struct folio *folio, bool silent) argument
462 __nilfs_clear_folio_dirty(struct folio *folio) argument
501 struct folio *folio; local
[all...]

Completed in 220 milliseconds

1234567891011>>