/linux-master/fs/netfs/ |
H A D | output.c | 406 iov_iter_xarray(&wreq->iter, ITER_SOURCE, &wreq->mapping->i_pages, wreq->start, 0);
|
/linux-master/fs/smb/client/ |
H A D | file.c | 49 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 77 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 110 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 143 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); 2922 iov_iter_xarray(&wdata->iter, ITER_SOURCE, &mapping->i_pages, 3012 /* At this point we hold neither the i_pages lock nor the page lock: 3073 XA_STATE(xas, &mapping->i_pages, *_start / PAGE_SIZE); 4574 XA_STATE(xas, &mapping->i_pages, first); 4591 XA_STATE(xas, &rdata->mapping->i_pages, rdata->offset / PAGE_SIZE); 4754 iov_iter_xarray(&rdata->iter, ITER_DEST, &rdata->mapping->i_pages, [all...] |
/linux-master/fs/ |
H A D | fs-writeback.c | 402 XA_STATE(xas, &mapping->i_pages, 0); 407 xa_lock_irq(&mapping->i_pages); 485 xa_unlock_irq(&mapping->i_pages); 511 * synchronizing against the i_pages lock. 513 * Grabbing old_wb->list_lock, inode->i_lock and the i_pages lock
|
H A D | inode.c | 399 xa_init_flags(&mapping->i_pages, XA_FLAGS_LOCK_IRQ | XA_FLAGS_ACCOUNT); 609 * We have to cycle the i_pages lock here because reclaim can be in the 613 xa_lock_irq(&inode->i_data.i_pages); 623 xa_unlock_irq(&inode->i_data.i_pages);
|
/linux-master/fs/afs/ |
H A D | file.c | 342 iov_iter_xarray(&fsreq->def_iter, ITER_DEST, &folio->mapping->i_pages,
|
H A D | dir.c | 117 XA_STATE(xas, &mapping->i_pages, 0); 192 XA_STATE(xas, &mapping->i_pages, 0); 226 XA_STATE(xas, &mapping->i_pages, 0); 312 iov_iter_xarray(&req->def_iter, ITER_DEST, &dvnode->netfs.inode.i_mapping->i_pages,
|
/linux-master/mm/ |
H A D | huge_memory.c | 2908 xa_lock(&swap_cache->i_pages); 2930 __xa_store(&folio->mapping->i_pages, head[i].index, 2933 __xa_store(&swap_cache->i_pages, offset + i, 2955 xa_unlock(&swap_cache->i_pages); 2962 xa_unlock(&folio->mapping->i_pages); 3045 XA_STATE_ORDER(xas, &folio->mapping->i_pages, folio->index, new_order);
|
H A D | migrate.c | 406 XA_STATE(xas, &mapping->i_pages, folio_index(folio)); 533 XA_STATE(xas, &mapping->i_pages, folio_index(src));
|
H A D | madvise.c | 224 XA_STATE(xas, &mapping->i_pages, linear_page_index(vma, start));
|
H A D | khugepaged.c | 1806 XA_STATE_ORDER(xas, &mapping->i_pages, start, HPAGE_PMD_ORDER); 1925 * The page must be locked, so we can drop the i_pages lock 2228 XA_STATE(xas, &mapping->i_pages, start);
|
H A D | vmscan.c | 695 xa_lock_irq(&mapping->i_pages); 719 * and thus under the i_pages lock, then this ordering is not required. 737 xa_unlock_irq(&mapping->i_pages); 763 xa_unlock_irq(&mapping->i_pages); 775 xa_unlock_irq(&mapping->i_pages);
|
/linux-master/include/linux/ |
H A D | fs.h | 444 * @i_pages: Cached pages. 454 * @nrpages: Number of page entries, protected by the i_pages lock. 465 struct xarray i_pages; member in struct:address_space 500 return xa_marked(&mapping->i_pages, tag); 2358 * wb stat updates to grab the i_pages lock. See
|
/linux-master/fs/ceph/ |
H A D | addr.c | 315 iov_iter_xarray(&iter, ITER_DEST, &rreq->mapping->i_pages, subreq->start, len); 372 iov_iter_xarray(&iter, ITER_DEST, &rreq->mapping->i_pages, subreq->start, len);
|
/linux-master/fs/btrfs/ |
H A D | compression.c | 454 page = xa_load(&mapping->i_pages, pg_index);
|
H A D | extent_io.c | 2108 * At this point we hold neither the i_pages lock nor 4105 xa_lock_irq(&folio->mapping->i_pages); 4107 __xa_clear_mark(&folio->mapping->i_pages, 4109 xa_unlock_irq(&folio->mapping->i_pages);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io-buffered.c | 120 folio = xa_load(&iter->mapping->i_pages, folio_offset);
|
/linux-master/fs/f2fs/ |
H A D | data.c | 4088 xa_lock_irqsave(&mapping->i_pages, flags); 4089 __xa_clear_mark(&mapping->i_pages, page_index(page), 4091 xa_unlock_irqrestore(&mapping->i_pages, flags);
|
H A D | node.c | 1415 apage = xa_load(&NODE_MAPPING(sbi)->i_pages, nid);
|
H A D | file.c | 417 xa_get_mark(&mapping->i_pages, index, PAGECACHE_TAG_DIRTY))
|