/linux-master/fs/fuse/ |
H A D | readdir.c | 38 struct page *page; local 57 /* Dirent doesn't fit in current page? Jump to next page. */ 65 page = find_lock_page(file->f_mapping, index); 67 page = find_or_create_page(file->f_mapping, index, 70 if (!page) 79 addr = kmap_local_page(page); 82 SetPageUptodate(page); 90 unlock_page(page); 334 struct page *page; local 453 struct page *page; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shmem.c | 38 struct page *page; local 43 for_each_sgt_page(page, sgt_iter, st) { 44 struct folio *folio = page_folio(page); 191 * shmemfs first checks if there is enough memory to allocate the page 213 struct page *page; local 242 for_each_sgt_page(page, sgt_iter, st) 243 put_page(page); 270 * shmemfs first checks if there is enough memory to allocate the page 323 struct page *page; local 465 struct page *page; local 678 struct page *page; local [all...] |
/linux-master/mm/ |
H A D | sparse.c | 36 * If we did not store the node number in the page then we have to 38 * node the page belongs to. 46 int page_to_nid(const struct page *page) argument 48 return section_to_node_table[page_to_section(page)]; 277 * physical page frame number. 279 static unsigned long sparse_encode_mem_map(struct page *mem_map, unsigned long pnum) 292 struct page *sparse_decode_mem_map(unsigned long coded_mem_map, unsigned long pnum) 296 return ((struct page *)coded_mem_map) + section_nr_to_pfn(pnum); 301 unsigned long pnum, struct page *mem_ma 723 struct page *page = virt_to_page(memmap); local [all...] |
H A D | secretmem.c | 57 struct page *page; local 68 page = find_lock_page(mapping, offset); 69 if (!page) { 76 page = &folio->page; 77 err = set_direct_map_invalid_noflush(page); 89 * If a split of large page was required, it 90 * already happened when we marked the page invalid 93 set_direct_map_default_noflush(page); [all...] |
H A D | hugetlb_vmemmap.h | 14 * Reserve one vmemmap page, all vmemmap addresses are mapped to it. See 18 #define HUGETLB_VMEMMAP_RESERVE_PAGES (HUGETLB_VMEMMAP_RESERVE_SIZE / sizeof(struct page)) 30 return pages_per_huge_page(h) * sizeof(struct page); 34 * Return how many vmemmap size associated with a HugeTLB page that can be 41 if (!is_power_of_2(sizeof(struct page)))
|
/linux-master/drivers/md/ |
H A D | md-bitmap.c | 41 * check a page and, if necessary, allocate it (or hijack it if the alloc fails) 43 * 1) check to see if this page is allocated, if it's not then try to alloc 44 * 2) if the alloc fails, set the page's hijacked flag so we'll use the 45 * page pointer directly as a counter 47 * if we find our page, we increment the page's refcount so that it stays 51 unsigned long page, int create, int no_hijack) 57 WARN_ON_ONCE(page >= bitmap->pages); 58 if (bitmap->bp[page].hijacked) /* it's hijacked, don't try to alloc */ 61 if (bitmap->bp[page] 112 md_bitmap_checkfree(struct bitmap_counts *bitmap, unsigned long page) argument 144 read_sb_page(struct mddev *mddev, loff_t offset, struct page *page, unsigned long index, int size) argument 224 __write_sb_page(struct md_rdev *rdev, struct bitmap *bitmap, unsigned long pg_index, struct page *page) argument 280 write_sb_page(struct bitmap *bitmap, unsigned long pg_index, struct page *page, bool wait) argument 300 write_file_page(struct bitmap *bitmap, struct page *page, int wait) argument 327 free_buffers(struct page *page) argument 351 read_file_page(struct file *file, unsigned long index, struct bitmap *bitmap, unsigned long count, struct page *page) argument 414 write_file_page(struct bitmap *bitmap, struct page *page, int wait) argument 417 read_file_page(struct file *file, unsigned long index, struct bitmap *bitmap, unsigned long count, struct page *page) argument 422 free_buffers(struct page *page) argument 439 struct page *page = store->filemap[pg_index]; local 949 struct page *page; local 980 struct page *page; local 1010 struct page *page; local 1147 struct page *page = store->filemap[i]; local 1170 struct page *page = store->filemap[i]; local 1194 struct page *page = filemap_get_page(&bitmap->storage, i); local 1252 unsigned long page = chunk >> PAGE_COUNTER_SHIFT; local 1260 unsigned long page = chunk >> PAGE_COUNTER_SHIFT; local 1425 unsigned long page = chunk >> PAGE_COUNTER_SHIFT; variable 2232 unsigned long page; local 2333 location_show(struct mddev *mddev, char *page) argument 2445 space_show(struct mddev *mddev, char *page) argument 2478 timeout_show(struct mddev *mddev, char *page) argument 2522 backlog_show(struct mddev *mddev, char *page) argument 2582 chunksize_show(struct mddev *mddev, char *page) argument 2611 metadata_show(struct mddev *mddev, char *page) argument 2638 can_clear_show(struct mddev *mddev, char *page) argument 2670 behind_writes_used_show(struct mddev *mddev, char *page) argument [all...] |
/linux-master/arch/sh/mm/ |
H A D | tlb-pteaex.c | 70 void local_flush_tlb_one(unsigned long asid, unsigned long page) argument 73 __raw_writel(page, MMU_UTLB_ADDRESS_ARRAY | MMU_PAGE_ASSOC_BIT); 75 __raw_writel(page, MMU_ITLB_ADDRESS_ARRAY | MMU_PAGE_ASSOC_BIT);
|
H A D | tlb-sh3.c | 55 void local_flush_tlb_one(unsigned long asid, unsigned long page) argument 66 addr = MMU_TLB_ADDRESS_ARRAY | (page & 0x1F000); 67 data = (page & 0xfffe0000) | asid; /* VALID bit is off */
|
/linux-master/arch/nios2/mm/ |
H A D | dma-mapping.c | 63 void arch_dma_prep_coherent(struct page *page, size_t size) argument 65 unsigned long start = (unsigned long)page_address(page);
|
/linux-master/arch/xtensa/include/asm/ |
H A D | highmem.h | 32 static inline int get_pkmap_color(struct page *page) argument 34 return DCACHE_ALIAS(page_to_phys(page));
|
/linux-master/arch/m68k/mm/ |
H A D | cache.c | 99 void flush_icache_user_page(struct vm_area_struct *vma, struct page *page, argument 117 : : "a" (page_to_phys(page)));
|
/linux-master/fs/xfs/ |
H A D | xfs_bio_io.c | 34 struct page *page = kmem_to_page(data); local 38 while (bio_add_page(bio, page, len, off) != len) {
|
/linux-master/fs/freevxfs/ |
H A D | vxfs_subr.c | 26 vxfs_put_page(struct page *pp) 33 * vxfs_get_page - read a page into memory. 35 * @n: page number 38 * vxfs_get_page reads the @n th page of @ip into the pagecache. 41 * The wanted page on success, else a NULL pointer. 43 struct page * 46 struct page * pp; 114 * vxfs_read_folio - read one page synchronously into the pagecache
|
/linux-master/scripts/gdb/linux/ |
H A D | page_owner.py | 22 t = """Usage: lx-dump-page-owner [Option] 26 lx-dump-page-owner --pfn 655360\n""" 31 """Dump page owner""" 39 super(DumpPageOwner, self).__init__("lx-dump-page-owner", gdb.COMMAND_SUPPORT) 79 def lookup_page_ext(self, page): 80 pfn = self.p_ops.page_to_pfn(page) 87 def page_ext_get(self, page): 88 page_ext = self.lookup_page_ext(page) 99 page = gdb.Value(struct_page_addr).cast(utils.get_page_type().pointer()) 100 pfn = self.p_ops.page_to_pfn(page) [all...] |
/linux-master/include/linux/ |
H A D | cma.h | 17 * -- can deal with only some pageblocks of a higher-order page being 47 extern struct page *cma_alloc(struct cma *cma, unsigned long count, unsigned int align, 49 extern bool cma_pages_valid(struct cma *cma, const struct page *pages, unsigned long count); 50 extern bool cma_release(struct cma *cma, const struct page *pages, unsigned long count);
|
/linux-master/tools/testing/selftests/exec/ |
H A D | Makefile | 32 $(CC) $(CFLAGS) $(LDFLAGS) -Wl,-z,max-page-size=0x1000 -pie -static $< -o $@ 34 $(CC) $(CFLAGS) $(LDFLAGS) -Wl,-z,max-page-size=0x200000 -pie -static $< -o $@ 36 $(CC) $(CFLAGS) $(LDFLAGS) -Wl,-z,max-page-size=0x1000000 -pie -static $< -o $@
|
/linux-master/arch/arm64/mm/ |
H A D | dma-mapping.c | 34 void arch_dma_prep_coherent(struct page *page, size_t size) argument 36 unsigned long start = (unsigned long)page_address(page);
|
/linux-master/arch/riscv/include/asm/ |
H A D | cacheflush.h | 32 static inline void flush_dcache_page(struct page *page) argument 34 flush_dcache_folio(page_folio(page));
|
/linux-master/drivers/net/ethernet/microchip/lan966x/ |
H A D | lan966x_fdma.c | 14 static struct page *lan966x_fdma_rx_alloc_page(struct lan966x_rx *rx, 17 struct page *page; local 19 page = page_pool_dev_alloc_pages(rx->page_pool); 20 if (unlikely(!page)) 23 db->dataptr = page_pool_get_dma_addr(page) + XDP_PACKET_HEADROOM; 25 return page; 35 rx->page[i][j], false); 41 struct page *page; local 109 struct page *page; local 469 struct page *page; local 499 struct page *page; local 554 struct page *page; local 712 struct page *page; local [all...] |
/linux-master/drivers/virtio/ |
H A D | virtio_balloon.c | 23 * Balloon device works in 4K page units. So each page is pointed to by 25 * page units. 35 /* The order of free page blocks to report to host */ 37 /* The size of a free page block in bytes */ 61 /* The free page reporting work item submitted to the balloon wq */ 77 /* The number of free page blocks on the above list */ 98 * Each page on this list adds VIRTIO_BALLOON_PAGES_PER_PAGE 119 /* Free page reporting device */ 137 static u32 page_to_balloon_pfn(struct page *pag argument 226 set_page_pfns(struct virtio_balloon *vb, __virtio32 pfns[], struct page *page) argument 246 struct page *page; local 254 struct page *page = balloon_page_alloc(); local 295 struct page *page, *next; local 309 struct page *page; local 450 struct page *page; local 673 struct page *page; local 802 virtballoon_migratepage(struct balloon_dev_info *vb_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) argument [all...] |
/linux-master/drivers/target/tcm_fc/ |
H A D | tfc_io.c | 57 struct page *page = NULL; local 84 page = sg_page(sg); 102 page = sg_page(sg); 131 BUG_ON(!page); 132 get_page(page); 135 page, off_in_page, tlen); 138 fp_skb(fp)->truesize += page_size(page); 140 BUG_ON(!page); 141 from = kmap_atomic(page 208 struct page *page = NULL; local [all...] |
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | scatterlist.c | 52 struct page *page = sg_page(sg); local 55 if (page_to_pfn(page) != pfn) { 57 __func__, who, pfn, page_to_pfn(page)); 90 struct page *page = sg_page_iter_page(&sgiter); local 92 if (page != pfn_to_page(pfn)) { 94 __func__, who, pfn, page_to_pfn(page)); 117 struct page *page; local [all...] |
/linux-master/fs/qnx6/ |
H A D | qnx6.h | 10 * 16-02-2012 page map extension by Al Viro 129 static inline void qnx6_put_page(struct page *page) argument 131 kunmap(page); 132 put_page(page); 136 struct page **res_page);
|
/linux-master/arch/arm/kernel/ |
H A D | patch.c | 26 struct page *page; local 29 page = vmalloc_to_page(addr); 31 page = virt_to_page(addr); 38 set_fixmap(fixmap, page_to_phys(page));
|
/linux-master/arch/parisc/kernel/ |
H A D | patch.c | 31 struct page *page; local 35 page = vmalloc_to_page(addr); 37 page = virt_to_page(addr); 42 set_fixmap(fixmap, page_to_phys(page)); 76 * We're crossing a page boundary, so
|