/linux-master/arch/nios2/include/asm/ |
H A D | cacheflush.h | 54 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 55 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) 57 xa_lock_irqsave(&mapping->i_pages, flags) 59 xa_unlock_irqrestore(&mapping->i_pages, flags)
|
/linux-master/arch/parisc/include/asm/ |
H A D | cacheflush.h | 55 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 56 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) 58 xa_lock_irqsave(&mapping->i_pages, flags) 60 xa_unlock_irqrestore(&mapping->i_pages, flags)
|
/linux-master/fs/nilfs2/ |
H A D | btnode.c | 198 xa_lock_irq(&btnc->i_pages); 199 err = __xa_insert(&btnc->i_pages, newkey, ofolio, GFP_NOFS); 200 xa_unlock_irq(&btnc->i_pages); 255 xa_lock_irq(&btnc->i_pages); 256 __xa_erase(&btnc->i_pages, oldkey); 257 __xa_set_mark(&btnc->i_pages, newkey, PAGECACHE_TAG_DIRTY); 258 xa_unlock_irq(&btnc->i_pages); 286 xa_erase_irq(&btnc->i_pages, newkey);
|
H A D | page.c | 328 xa_lock_irq(&smap->i_pages); 329 f = __xa_erase(&smap->i_pages, index); 332 xa_unlock_irq(&smap->i_pages); 334 xa_lock_irq(&dmap->i_pages); 335 f = __xa_store(&dmap->i_pages, index, folio, GFP_NOFS); 344 __xa_set_mark(&dmap->i_pages, index, 347 xa_unlock_irq(&dmap->i_pages); 467 xa_lock_irq(&mapping->i_pages); 469 __xa_clear_mark(&mapping->i_pages, folio->index, 471 xa_unlock_irq(&mapping->i_pages); [all...] |
/linux-master/arch/csky/abiv1/inc/abi/ |
H A D | cacheflush.h | 19 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 20 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/linux-master/mm/ |
H A D | truncate.c | 34 XA_STATE(xas, &mapping->i_pages, index); 46 xa_lock_irq(&mapping->i_pages); 48 xa_unlock_irq(&mapping->i_pages); 79 xa_lock_irq(&mapping->i_pages); 100 xa_unlock_irq(&mapping->i_pages); 470 xa_lock_irq(&mapping->i_pages); 471 xa_unlock_irq(&mapping->i_pages); 570 xa_lock_irq(&mapping->i_pages); 576 xa_unlock_irq(&mapping->i_pages); 584 xa_unlock_irq(&mapping->i_pages); [all...] |
H A D | workingset.c | 379 * Return: a shadow entry to be stored in @folio->mapping->i_pages in place 628 * as node->private_list is protected by the i_pages lock. 630 mapping = container_of(node->array, struct address_space, i_pages); 631 lockdep_assert_held(&mapping->i_pages.xa_lock); 715 * the shadow node LRU under the i_pages lock and the 720 * We can then safely transition to the i_pages lock to 725 mapping = container_of(node->array, struct address_space, i_pages); 728 if (!xa_trylock(&mapping->i_pages)) { 737 xa_unlock(&mapping->i_pages); 762 xa_unlock_irq(&mapping->i_pages); [all...] |
H A D | swap_state.c | 78 page = xa_load(&address_space->i_pages, idx); 93 XA_STATE_ORDER(xas, &address_space->i_pages, idx, folio_order(folio)); 148 XA_STATE(xas, &address_space->i_pages, idx); 240 xa_lock_irq(&address_space->i_pages); 242 xa_unlock_irq(&address_space->i_pages); 257 XA_STATE(xas, &address_space->i_pages, curr); 261 xa_lock_irq(&address_space->i_pages); 267 xa_unlock_irq(&address_space->i_pages); 717 xa_init_flags(&space->i_pages, XA_FLAGS_LOCK_IRQ);
|
H A D | filemap.c | 83 * ->i_pages lock 92 * ->i_pages lock (arch-dependent flush_dcache_mmap_lock) 103 * ->i_pages lock (__sync_single_inode) 114 * ->i_pages lock (try_to_unmap_one) 118 * ->i_pages lock (folio_remove_rmap_pte->set_page_dirty) 139 XA_STATE(xas, &mapping->i_pages, folio->index); 224 * is safe. The caller must hold the i_pages lock. 263 xa_lock_irq(&mapping->i_pages); 265 xa_unlock_irq(&mapping->i_pages); 278 * The function walks over mapping->i_pages an [all...] |
H A D | readahead.c | 230 struct folio *folio = xa_load(&mapping->i_pages, index + i); 784 struct folio *folio = xa_load(&mapping->i_pages, index); 811 struct folio *folio = xa_load(&mapping->i_pages, index);
|
H A D | page-writeback.c | 2344 XA_STATE(xas, &mapping->i_pages, start); 2720 xa_lock_irqsave(&mapping->i_pages, flags); 2724 __xa_set_mark(&mapping->i_pages, folio_index(folio), 2727 xa_unlock_irqrestore(&mapping->i_pages, flags); 3009 xa_lock_irqsave(&mapping->i_pages, flags); 3011 __xa_clear_mark(&mapping->i_pages, folio_index(folio), 3026 xa_unlock_irqrestore(&mapping->i_pages, flags); 3049 XA_STATE(xas, &mapping->i_pages, folio_index(folio));
|
H A D | memfd.c | 73 XA_STATE(xas, &mapping->i_pages, 0);
|
H A D | shmem.c | 460 xa_lock_irq(&mapping->i_pages); 462 xa_unlock_irq(&mapping->i_pages); 482 XA_STATE(xas, &mapping->i_pages, index); 504 return xa_load(&mapping->i_pages, index) == swp_to_radix_entry(swap); 771 XA_STATE_ORDER(xas, &mapping->i_pages, index, folio_order(folio)); 826 xa_lock_irq(&mapping->i_pages); 832 xa_unlock_irq(&mapping->i_pages); 845 old = xa_cmpxchg_irq(&mapping->i_pages, index, radswap, NULL, 0); 856 * This is safe to call without i_rwsem or the i_pages lock thanks to RCU, 862 XA_STATE(xas, &mapping->i_pages, star [all...] |
/linux-master/include/linux/ |
H A D | backing-dev.h | 245 * holding either @inode->i_lock, the i_pages lock, or the 253 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && 276 * holding inode->i_lock, the i_pages lock or wb->list_lock. This 297 xa_lock_irqsave(&inode->i_mapping->i_pages, cookie->flags); 300 * Protected by either !I_WB_SWITCH + rcu_read_lock() or the i_pages 315 xa_unlock_irqrestore(&inode->i_mapping->i_pages, cookie->flags);
|
H A D | pagemap.h | 141 return xa_empty(&mapping->i_pages); 152 * hold the i_pages lock, which usually protects cache state. That's 154 * its LRU state don't nest inside the irq-safe i_pages lock. 178 head = rcu_access_pointer(mapping->i_pages.xa_head); 1339 folio = xa_load(&ractl->mapping->i_pages, ractl->_index); 1383 XA_STATE(xas, &rac->mapping->i_pages, 0);
|
/linux-master/fs/netfs/ |
H A D | buffered_read.c | 26 XA_STATE(xas, &rreq->mapping->i_pages, start_page); 214 iov_iter_xarray(&rreq->iter, ITER_DEST, &ractl->mapping->i_pages, 314 iov_iter_xarray(&rreq->iter, ITER_DEST, &mapping->i_pages, 480 iov_iter_xarray(&rreq->iter, ITER_DEST, &mapping->i_pages, 548 iov_iter_xarray(&rreq->iter, ITER_DEST, &mapping->i_pages,
|
H A D | fscache_io.c | 181 XA_STATE(xas, &mapping->i_pages, first); 251 iov_iter_xarray(&iter, ITER_SOURCE, &mapping->i_pages, start, len);
|
H A D | io.c | 118 XA_STATE(xas, &rreq->mapping->i_pages, subreq->start / PAGE_SIZE); 209 iov_iter_xarray(&iter, ITER_SOURCE, &rreq->mapping->i_pages,
|
H A D | buffered_write.c | 645 XA_STATE(xas, &mapping->i_pages, wreq->start / PAGE_SIZE); 974 iov_iter_xarray(&wreq->iter, ITER_SOURCE, &mapping->i_pages, start, 1056 /* At this point we hold neither the i_pages lock nor the page lock: 1120 XA_STATE(xas, &mapping->i_pages, *_start / PAGE_SIZE);
|
/linux-master/arch/arm/include/asm/ |
H A D | cacheflush.h | 321 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 322 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/linux-master/fs/erofs/ |
H A D | fscache.c | 62 XA_STATE(xas, &req->mapping->i_pages, start_page); 243 iov_iter_xarray(&io->iter, ITER_DEST, &folio->mapping->i_pages, 289 iov_iter_xarray(&iter, ITER_DEST, &mapping->i_pages, pos, PAGE_SIZE); 304 iov_iter_xarray(&iter, ITER_DEST, &mapping->i_pages, pos, count); 324 iov_iter_xarray(&io->iter, ITER_DEST, &mapping->i_pages, pos, count);
|
/linux-master/fs/ |
H A D | dax.c | 191 * under the i_pages lock, ditto for entry handling in our callers. 204 * If @order is larger than the order of the entry found in i_pages, this 207 * Must be called with the i_pages lock held. 239 * The only thing keeping the address space around is the i_pages lock 240 * (it's cycled in clear_inode() after removing the entries from i_pages) 448 xas.xa = &mapping->i_pages; 473 XA_STATE(xas, &mapping->i_pages, folio->index); 502 xas.xa = &mapping->i_pages; 536 XA_STATE(xas, &mapping->i_pages, index); 605 * the i_pages loc [all...] |
/linux-master/fs/orangefs/ |
H A D | inode.c | 246 struct xarray *i_pages; local 264 i_pages = &rac->mapping->i_pages; 266 iov_iter_xarray(&iter, ITER_DEST, i_pages, offset, readahead_length(rac));
|
/linux-master/fs/smb/client/ |
H A D | fscache.c | 177 iov_iter_xarray(&iter, ITER_SOURCE, &inode->i_mapping->i_pages, start, len);
|
/linux-master/fs/nfs/ |
H A D | fscache.c | 319 xa_for_each_range(&sreq->rreq->mapping->i_pages, idx, page, start, last) {
|