/linux-master/fs/smb/client/ |
H A D | fscache.c | 211 iov_iter_xarray(&iter, ITER_SOURCE, &inode->i_mapping->i_pages, start, len);
|
H A D | file.c | 49 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 77 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 110 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 143 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 2922 iov_iter_xarray(&wdata->iter, ITER_SOURCE, &mapping->i_pages, 3012 /* At this point we hold neither the i_pages lock nor the page lock: 3073 XA_STATE(xas, &mapping->i_pages, *_start / PAGE_SIZE); 4574 XA_STATE(xas, &mapping->i_pages, first); 4591 XA_STATE(xas, &rdata->mapping->i_pages, rdata->offset / PAGE_SIZE); 4754 iov_iter_xarray(&rdata->iter, ITER_DEST, &rdata->mapping->i_pages, [all...] |
/linux-master/mm/ |
H A D | shmem.c | 460 xa_lock_irq(&mapping->i_pages); 462 xa_unlock_irq(&mapping->i_pages); 482 XA_STATE(xas, &mapping->i_pages, index); 504 return xa_load(&mapping->i_pages, index) == swp_to_radix_entry(swap); 765 XA_STATE_ORDER(xas, &mapping->i_pages, index, folio_order(folio)); 820 xa_lock_irq(&mapping->i_pages); 826 xa_unlock_irq(&mapping->i_pages); 839 old = xa_cmpxchg_irq(&mapping->i_pages, index, radswap, NULL, 0); 850 * This is safe to call without i_rwsem or the i_pages lock thanks to RCU, 856 XA_STATE(xas, &mapping->i_pages, star [all...] |
H A D | madvise.c | 224 XA_STATE(xas, &mapping->i_pages, linear_page_index(vma, start));
|
H A D | huge_memory.c | 2908 xa_lock(&swap_cache->i_pages); 2930 __xa_store(&folio->mapping->i_pages, head[i].index, 2933 __xa_store(&swap_cache->i_pages, offset + i, 2955 xa_unlock(&swap_cache->i_pages); 2962 xa_unlock(&folio->mapping->i_pages); 3045 XA_STATE_ORDER(xas, &folio->mapping->i_pages, folio->index, new_order);
|
H A D | filemap.c | 83 * ->i_pages lock 92 * ->i_pages lock (arch-dependent flush_dcache_mmap_lock) 103 * ->i_pages lock (__sync_single_inode) 114 * ->i_pages lock (try_to_unmap_one) 118 * ->i_pages lock (folio_remove_rmap_pte->set_page_dirty) 139 XA_STATE(xas, &mapping->i_pages, folio->index); 224 * is safe. The caller must hold the i_pages lock. 263 xa_lock_irq(&mapping->i_pages); 265 xa_unlock_irq(&mapping->i_pages); 278 * The function walks over mapping->i_pages an [all...] |
H A D | workingset.c | 379 * Return: a shadow entry to be stored in @folio->mapping->i_pages in place 628 * as node->private_list is protected by the i_pages lock. 630 mapping = container_of(node->array, struct address_space, i_pages); 631 lockdep_assert_held(&mapping->i_pages.xa_lock); 715 * the shadow node LRU under the i_pages lock and the 720 * We can then safely transition to the i_pages lock to 725 mapping = container_of(node->array, struct address_space, i_pages); 728 if (!xa_trylock(&mapping->i_pages)) { 737 xa_unlock(&mapping->i_pages); 762 xa_unlock_irq(&mapping->i_pages); [all...] |
H A D | vmscan.c | 695 xa_lock_irq(&mapping->i_pages); 719 * and thus under the i_pages lock, then this ordering is not required. 737 xa_unlock_irq(&mapping->i_pages); 763 xa_unlock_irq(&mapping->i_pages); 775 xa_unlock_irq(&mapping->i_pages);
|
H A D | swap_state.c | 78 page = xa_load(&address_space->i_pages, idx); 93 XA_STATE_ORDER(xas, &address_space->i_pages, idx, folio_order(folio)); 148 XA_STATE(xas, &address_space->i_pages, idx); 240 xa_lock_irq(&address_space->i_pages); 242 xa_unlock_irq(&address_space->i_pages); 257 XA_STATE(xas, &address_space->i_pages, curr); 261 xa_lock_irq(&address_space->i_pages); 267 xa_unlock_irq(&address_space->i_pages); 717 xa_init_flags(&space->i_pages, XA_FLAGS_LOCK_IRQ);
|
H A D | readahead.c | 230 struct folio *folio = xa_load(&mapping->i_pages, index + i); 784 struct folio *folio = xa_load(&mapping->i_pages, index); 811 struct folio *folio = xa_load(&mapping->i_pages, index);
|
H A D | page-writeback.c | 2344 XA_STATE(xas, &mapping->i_pages, start); 2720 xa_lock_irqsave(&mapping->i_pages, flags); 2724 __xa_set_mark(&mapping->i_pages, folio_index(folio), 2727 xa_unlock_irqrestore(&mapping->i_pages, flags); 3009 xa_lock_irqsave(&mapping->i_pages, flags); 3011 __xa_clear_mark(&mapping->i_pages, folio_index(folio), 3026 xa_unlock_irqrestore(&mapping->i_pages, flags); 3049 XA_STATE(xas, &mapping->i_pages, folio_index(folio));
|
H A D | migrate.c | 406 XA_STATE(xas, &mapping->i_pages, folio_index(folio)); 533 XA_STATE(xas, &mapping->i_pages, folio_index(src));
|
H A D | memfd.c | 73 XA_STATE(xas, &mapping->i_pages, 0);
|
H A D | khugepaged.c | 1806 XA_STATE_ORDER(xas, &mapping->i_pages, start, HPAGE_PMD_ORDER); 1925 * The page must be locked, so we can drop the i_pages lock 2228 XA_STATE(xas, &mapping->i_pages, start);
|
/linux-master/fs/btrfs/ |
H A D | extent_io.c | 2108 * At this point we hold neither the i_pages lock nor 4105 xa_lock_irq(&folio->mapping->i_pages); 4107 __xa_clear_mark(&folio->mapping->i_pages, 4109 xa_unlock_irq(&folio->mapping->i_pages);
|
/linux-master/fs/ceph/ |
H A D | addr.c | 315 iov_iter_xarray(&iter, ITER_DEST, &rreq->mapping->i_pages, subreq->start, len); 372 iov_iter_xarray(&iter, ITER_DEST, &rreq->mapping->i_pages, subreq->start, len);
|
/linux-master/include/linux/ |
H A D | fs.h | 444 * @i_pages: Cached pages. 454 * @nrpages: Number of page entries, protected by the i_pages lock. 465 struct xarray i_pages; member in struct:address_space 500 return xa_marked(&mapping->i_pages, tag); 2358 * wb stat updates to grab the i_pages lock. See
|
/linux-master/fs/netfs/ |
H A D | fscache_io.c | 181 XA_STATE(xas, &mapping->i_pages, first); 251 iov_iter_xarray(&iter, ITER_SOURCE, &mapping->i_pages, start, len);
|
/linux-master/fs/f2fs/ |
H A D | node.c | 1415 apage = xa_load(&NODE_MAPPING(sbi)->i_pages, nid);
|
H A D | file.c | 417 xa_get_mark(&mapping->i_pages, index, PAGECACHE_TAG_DIRTY))
|
H A D | data.c | 4088 xa_lock_irqsave(&mapping->i_pages, flags); 4089 __xa_clear_mark(&mapping->i_pages, page_index(page), 4091 xa_unlock_irqrestore(&mapping->i_pages, flags);
|
/linux-master/fs/afs/ |
H A D | dir.c | 117 XA_STATE(xas, &mapping->i_pages, 0); 192 XA_STATE(xas, &mapping->i_pages, 0); 226 XA_STATE(xas, &mapping->i_pages, 0); 312 iov_iter_xarray(&req->def_iter, ITER_DEST, &dvnode->netfs.inode.i_mapping->i_pages,
|
/linux-master/fs/nfs/ |
H A D | fscache.c | 319 xa_for_each_range(&sreq->rreq->mapping->i_pages, idx, page, start, last) {
|
/linux-master/fs/ |
H A D | inode.c | 399 xa_init_flags(&mapping->i_pages, XA_FLAGS_LOCK_IRQ | XA_FLAGS_ACCOUNT); 609 * We have to cycle the i_pages lock here because reclaim can be in the 613 xa_lock_irq(&inode->i_data.i_pages); 623 xa_unlock_irq(&inode->i_data.i_pages);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io-buffered.c | 120 folio = xa_load(&iter->mapping->i_pages, folio_offset);
|