Lines Matching refs:page
32 * - Tracks whether the page is UC, WB or cached (and reverts to WB
59 #define NUM_PAGES_TO_ALLOC (PAGE_SIZE/sizeof(struct page *))
123 * The accounting page keeping track of the allocated page along with
126 * @vaddr: The virtual address of the page
127 * @dma: The bus address of the page. If the page is not allocated
133 struct page *p;
274 static int set_pages_array_wb(struct page **pages, int addrinarray)
285 static int set_pages_array_wc(struct page **pages, int addrinarray)
296 static int set_pages_array_uc(struct page **pages, int addrinarray)
309 struct page **pages, unsigned cpages)
312 /* Set page caching */
381 struct page *pages[], unsigned npages)
385 /* Don't set WB on WB page pool. */
399 /* Don't set WB on WB page pool. */
421 struct page **pages_to_free;
435 pages_to_free = kmalloc(npages_to_free * sizeof(struct page *),
462 * Because changing page caching is costly
676 struct page **failed_pages,
680 struct page *p;
686 /* Find the failed page. */
711 struct page **caching_array;
713 struct page *p;
717 (unsigned)(PAGE_SIZE/sizeof(struct page *)));
719 /* allocate array for page caching change */
720 caching_array = kmalloc(max_cpages*sizeof(struct page *), GFP_KERNEL);
736 pr_err("%s: Unable to get page %u\n",
754 /* gfp flags of highmem page should never be dma32 so we
836 * allocates one page at a time.
1007 * Callback for mm to request pool to reduce number of page held.