• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/mm/

Lines Matching refs:mapping

42 	invalidatepage = page->mapping->a_ops->invalidatepage;
75 struct address_space *mapping = page->mapping;
76 if (mapping && mapping_cap_account_dirty(mapping)) {
78 dec_bdi_stat(mapping->backing_dev_info,
92 * We need to bale out if page->mapping is no longer equal to the original
93 * mapping. This happens a) when the VM reclaimed the page while we waited on
98 truncate_complete_page(struct address_space *mapping, struct page *page)
100 if (page->mapping != mapping)
124 invalidate_complete_page(struct address_space *mapping, struct page *page)
128 if (page->mapping != mapping)
135 ret = remove_mapping(mapping, page);
140 int truncate_inode_page(struct address_space *mapping, struct page *page)
143 unmap_mapping_range(mapping,
147 return truncate_complete_page(mapping, page);
153 int generic_error_remove_page(struct address_space *mapping, struct page *page)
155 if (!mapping)
161 if (!S_ISREG(mapping->host->i_mode))
163 return truncate_inode_page(mapping, page);
168 * Safely invalidate one page from its pagecache mapping.
175 struct address_space *mapping = page_mapping(page);
176 if (!mapping)
182 return invalidate_complete_page(mapping, page);
187 * @mapping: mapping to truncate
205 * mapping is large, it is probably the case that the final pages are the most
208 void truncate_inode_pages_range(struct address_space *mapping,
218 if (mapping->nrpages == 0)
227 pagevec_lookup(&pvec, mapping, next, PAGEVEC_SIZE)) {
246 truncate_inode_page(mapping, page);
254 struct page *page = find_lock_page(mapping, start - 1);
266 if (!pagevec_lookup(&pvec, mapping, next, PAGEVEC_SIZE)) {
284 truncate_inode_page(mapping, page);
298 * @mapping: mapping to truncate
303 void truncate_inode_pages(struct address_space *mapping, loff_t lstart)
305 truncate_inode_pages_range(mapping, lstart, (loff_t)-1);
311 * @mapping: the address_space which holds the pages to invalidate
322 unsigned long invalidate_mapping_pages(struct address_space *mapping,
332 pagevec_lookup(&pvec, mapping, next, PAGEVEC_SIZE)) {
376 invalidate_complete_page2(struct address_space *mapping, struct page *page)
378 if (page->mapping != mapping)
384 spin_lock_irq(&mapping->tree_lock);
391 spin_unlock_irq(&mapping->tree_lock);
396 spin_unlock_irq(&mapping->tree_lock);
400 static int do_launder_page(struct address_space *mapping, struct page *page)
404 if (page->mapping != mapping || mapping->a_ops->launder_page == NULL)
406 return mapping->a_ops->launder_page(page);
411 * @mapping: the address_space
420 int invalidate_inode_pages2_range(struct address_space *mapping,
434 pagevec_lookup(&pvec, mapping, next,
442 if (page->mapping != mapping) {
460 unmap_mapping_range(mapping,
470 unmap_mapping_range(mapping,
476 ret2 = do_launder_page(mapping, page);
478 if (!invalidate_complete_page2(mapping, page))
495 * @mapping: the address_space
502 int invalidate_inode_pages2(struct address_space *mapping)
504 return invalidate_inode_pages2_range(mapping, 0, -1);
526 struct address_space *mapping = inode->i_mapping;
537 unmap_mapping_range(mapping, new + PAGE_SIZE - 1, 0, 1);
538 truncate_inode_pages(mapping, new);
539 unmap_mapping_range(mapping, new + PAGE_SIZE - 1, 0, 1);