Lines Matching defs:pages
94 * BLOCKS_PER_PAGE on indirect pages, assume PAGE_CACHE_SIZE:
162 * pages are allocated, in order to allow huge sparse files.
172 static inline void shmem_unacct_blocks(unsigned long flags, long pages)
175 vm_unacct_memory(pages * VM_ACCT(PAGE_CACHE_SIZE));
195 static void shmem_free_blocks(struct inode *inode, long pages)
200 sbinfo->free_blocks += pages;
201 inode->i_blocks -= pages*BLOCKS_PER_PAGE;
212 * undirtied hole pages behind our back.
250 * For pages with index > SHMEM_NR_DIRECT there is the pointer
360 * page (and perhaps indirect index pages) yet to allocate:
698 * Empty swap vector directory pages to be freed?
733 * detect if any pages might have been added to cache
1039 lru_add_drain(); /* Push any new pages onto the LRU now */
1586 * Our dirty pages are not counted in nr_dirty,
1587 * and we do not attempt to balance dirty pages.