Searched refs:i_pages (Results 26 - 44 of 44) sorted by relevance

12

/linux-master/fs/netfs/
H A Doutput.c406 iov_iter_xarray(&wreq->iter, ITER_SOURCE, &wreq->mapping->i_pages, wreq->start, 0);
/linux-master/fs/smb/client/
H A Dfile.c49 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
77 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
110 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
143 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
2922 iov_iter_xarray(&wdata->iter, ITER_SOURCE, &mapping->i_pages,
3012 /* At this point we hold neither the i_pages lock nor the page lock:
3073 XA_STATE(xas, &mapping->i_pages, *_start / PAGE_SIZE);
4574 XA_STATE(xas, &mapping->i_pages, first);
4591 XA_STATE(xas, &rdata->mapping->i_pages, rdata->offset / PAGE_SIZE);
4754 iov_iter_xarray(&rdata->iter, ITER_DEST, &rdata->mapping->i_pages,
[all...]
/linux-master/fs/
H A Dfs-writeback.c402 XA_STATE(xas, &mapping->i_pages, 0);
407 xa_lock_irq(&mapping->i_pages);
485 xa_unlock_irq(&mapping->i_pages);
511 * synchronizing against the i_pages lock.
513 * Grabbing old_wb->list_lock, inode->i_lock and the i_pages lock
H A Dinode.c399 xa_init_flags(&mapping->i_pages, XA_FLAGS_LOCK_IRQ | XA_FLAGS_ACCOUNT);
609 * We have to cycle the i_pages lock here because reclaim can be in the
613 xa_lock_irq(&inode->i_data.i_pages);
623 xa_unlock_irq(&inode->i_data.i_pages);
/linux-master/fs/afs/
H A Dfile.c342 iov_iter_xarray(&fsreq->def_iter, ITER_DEST, &folio->mapping->i_pages,
H A Ddir.c117 XA_STATE(xas, &mapping->i_pages, 0);
192 XA_STATE(xas, &mapping->i_pages, 0);
226 XA_STATE(xas, &mapping->i_pages, 0);
312 iov_iter_xarray(&req->def_iter, ITER_DEST, &dvnode->netfs.inode.i_mapping->i_pages,
/linux-master/mm/
H A Dhuge_memory.c2908 xa_lock(&swap_cache->i_pages);
2930 __xa_store(&folio->mapping->i_pages, head[i].index,
2933 __xa_store(&swap_cache->i_pages, offset + i,
2955 xa_unlock(&swap_cache->i_pages);
2962 xa_unlock(&folio->mapping->i_pages);
3045 XA_STATE_ORDER(xas, &folio->mapping->i_pages, folio->index, new_order);
H A Dmigrate.c406 XA_STATE(xas, &mapping->i_pages, folio_index(folio));
533 XA_STATE(xas, &mapping->i_pages, folio_index(src));
H A Dmadvise.c224 XA_STATE(xas, &mapping->i_pages, linear_page_index(vma, start));
H A Dkhugepaged.c1806 XA_STATE_ORDER(xas, &mapping->i_pages, start, HPAGE_PMD_ORDER);
1925 * The page must be locked, so we can drop the i_pages lock
2228 XA_STATE(xas, &mapping->i_pages, start);
H A Dvmscan.c695 xa_lock_irq(&mapping->i_pages);
719 * and thus under the i_pages lock, then this ordering is not required.
737 xa_unlock_irq(&mapping->i_pages);
763 xa_unlock_irq(&mapping->i_pages);
775 xa_unlock_irq(&mapping->i_pages);
/linux-master/include/linux/
H A Dfs.h444 * @i_pages: Cached pages.
454 * @nrpages: Number of page entries, protected by the i_pages lock.
465 struct xarray i_pages; member in struct:address_space
500 return xa_marked(&mapping->i_pages, tag);
2358 * wb stat updates to grab the i_pages lock. See
/linux-master/fs/ceph/
H A Daddr.c315 iov_iter_xarray(&iter, ITER_DEST, &rreq->mapping->i_pages, subreq->start, len);
372 iov_iter_xarray(&iter, ITER_DEST, &rreq->mapping->i_pages, subreq->start, len);
/linux-master/fs/btrfs/
H A Dcompression.c454 page = xa_load(&mapping->i_pages, pg_index);
H A Dextent_io.c2108 * At this point we hold neither the i_pages lock nor
4105 xa_lock_irq(&folio->mapping->i_pages);
4107 __xa_clear_mark(&folio->mapping->i_pages,
4109 xa_unlock_irq(&folio->mapping->i_pages);
/linux-master/fs/bcachefs/
H A Dfs-io-buffered.c120 folio = xa_load(&iter->mapping->i_pages, folio_offset);
/linux-master/fs/f2fs/
H A Ddata.c4088 xa_lock_irqsave(&mapping->i_pages, flags);
4089 __xa_clear_mark(&mapping->i_pages, page_index(page),
4091 xa_unlock_irqrestore(&mapping->i_pages, flags);
H A Dnode.c1415 apage = xa_load(&NODE_MAPPING(sbi)->i_pages, nid);
H A Dfile.c417 xa_get_mark(&mapping->i_pages, index, PAGECACHE_TAG_DIRTY))

Completed in 375 milliseconds

12