/linux-master/arch/csky/include/asm/ |
H A D | tlbflush.h | 11 * - flush_tlb_page(vma, vmaddr) flushes one page 17 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
|
/linux-master/arch/s390/kernel/vdso32/ |
H A D | vdso32_wrapper.S | 4 #include <asm/page.h>
|
/linux-master/drivers/dma/ioat/ |
H A D | sysfs.c | 17 static ssize_t cap_show(struct dma_chan *c, char *page) argument 21 return sprintf(page, "copy%s%s%s%s%s\n", 31 static ssize_t version_show(struct dma_chan *c, char *page) argument 36 return sprintf(page, "%d.%d\n", 42 ioat_attr_show(struct kobject *kobj, struct attribute *attr, char *page) argument 52 return entry->show(&ioat_chan->dma_chan, page); 57 const char *page, size_t count) 67 return entry->store(&ioat_chan->dma_chan, page, count); 111 static ssize_t ring_size_show(struct dma_chan *c, char *page) argument 115 return sprintf(page, " 56 ioat_attr_store(struct kobject *kobj, struct attribute *attr, const char *page, size_t count) argument 119 ring_active_show(struct dma_chan *c, char *page) argument 128 intr_coalesce_show(struct dma_chan *c, char *page) argument 135 intr_coalesce_store(struct dma_chan *c, const char *page, size_t count) argument [all...] |
/linux-master/arch/parisc/kernel/vdso32/ |
H A D | vdso32_wrapper.S | 3 #include <asm/page.h>
|
/linux-master/arch/parisc/kernel/vdso64/ |
H A D | vdso64_wrapper.S | 3 #include <asm/page.h>
|
/linux-master/arch/powerpc/kernel/ |
H A D | vdso32_wrapper.S | 3 #include <asm/page.h>
|
H A D | vdso64_wrapper.S | 3 #include <asm/page.h>
|
/linux-master/arch/s390/include/asm/ |
H A D | user.h | 11 #include <asm/page.h> 22 UPAGE: 1 page consisting of a user struct that tells gdb what is present 26 always be only one page. 29 that may have been malloced. No attempt is made to determine if a page 30 is demand-zero or if a page is totally unused, we just cover the entire
|
/linux-master/arch/loongarch/vdso/ |
H A D | vdso.S | 11 #include <asm/page.h>
|
/linux-master/include/linux/ |
H A D | kernel-page-flags.h | 5 #include <uapi/linux/kernel-page-flags.h>
|
H A D | mmdebug.h | 8 struct page; 13 void dump_page(const struct page *page, const char *reason); 20 #define VM_BUG_ON_PAGE(cond, page) \ 23 dump_page(page, "VM_BUG_ON_PAGE(" __stringify(cond)")");\ 30 dump_page(&folio->page, "VM_BUG_ON_FOLIO(" __stringify(cond)")");\ 48 #define VM_WARN_ON_ONCE_PAGE(cond, page) ({ \ 53 dump_page(page, "VM_WARN_ON_ONCE_PAGE(" __stringify(cond)")");\ 63 dump_page(&folio->page, "VM_WARN_ON_FOLIO(" __stringify(cond)")");\ 73 dump_page(&folio->page, "VM_WARN_ON_ONCE_FOLI [all...] |
/linux-master/arch/x86/entry/vdso/vdso32/ |
H A D | vdso32.lds.S | 10 #include <asm/page.h>
|
/linux-master/arch/xtensa/include/asm/ |
H A D | tlb.h | 15 #include <asm/page.h>
|
/linux-master/arch/arm/include/asm/ |
H A D | kexec.h | 69 static inline unsigned long page_to_boot_pfn(struct page *page) argument 71 return page_to_pfn(page) + (arch_phys_to_idmap_offset >> PAGE_SHIFT); 75 static inline struct page *boot_pfn_to_page(unsigned long boot_pfn)
|
/linux-master/mm/ |
H A D | debug.c | 54 static void __dump_folio(struct folio *folio, struct page *page, argument 62 * page->_mapcount space in struct page is used by slab pages to 66 mapcount = atomic_read(&page->_mapcount) + 1; 71 pr_warn("page: refcount:%d mapcount:%d mapping:%p index:%#lx pfn:%#lx\n", 102 pr_warn("page_type: %pGt\n", &folio->page.page_type); 105 sizeof(unsigned long), page, 106 sizeof(struct page), false); 110 2 * sizeof(struct page), fals 113 __dump_page(const struct page *page) argument 152 dump_page(const struct page *page, const char *reason) argument 284 page_init_poison(struct page *page, size_t size) argument [all...] |
H A D | highmem.c | 13 * Rewrote high memory support to move the page cache into 62 * Determine color of virtual address where the page should be mapped. 64 static inline unsigned int get_pkmap_color(struct page *page) argument 71 * Get next index for mapping inside PKMAP region for page with given color. 82 * Determine if page index inside PKMAP region (pkmap_nr) of given color 154 struct page *__kmap_to_page(void *vaddr) 193 struct page *page; local 212 * no-one has the page mappe 234 map_new_virtual(struct page *page) argument 296 kmap_high(struct page *page) argument 326 kmap_high_get(struct page *page) argument 348 kunmap_high(struct page *page) argument 392 zero_user_segments(struct page *page, unsigned start1, unsigned end1, unsigned start2, unsigned end2) argument 500 arch_kmap_local_high_get(struct page *page) argument 564 __kmap_local_page_prot(struct page *page, pgprot_t prot) argument 716 struct page *page; member in struct:page_address_map 731 page_slot(const struct page *page) argument 742 page_address(const struct page *page) argument 775 set_page_address(struct page *page, void *virtual) argument [all...] |
/linux-master/fs/jfs/ |
H A D | jfs_metapage.h | 23 sector_t index; /* block address of page */ 27 struct page *page; member in struct:metapage 65 * hold_metapage and put_metapage are used in conjunction. The page lock 93 struct page *page = mp->page; local 94 lock_page(page); 97 get_page(page); 98 wait_on_page_writeback(page); [all...] |
/linux-master/include/xen/interface/io/ |
H A D | kbdif.h | 146 * page-gref 150 * a sole page in a single page sized event ring buffer. 152 * page-ref 156 * PFN of the shared page. 192 #define XENKBD_FIELD_RING_GREF "page-gref" 202 #define XENKBD_FIELD_RING_REF "page-ref" 210 * each other using a shared page and an event channel. 211 * Shared page contains a ring with event structures. 518 * Shared page [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | page_32.h | 20 static inline void clear_page(void *page) argument 22 memset(page, 0, PAGE_SIZE);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io-buffered.h | 14 unsigned, struct page **, void **); 16 unsigned, unsigned, struct page *, void *);
|
/linux-master/arch/mips/mm/ |
H A D | dma-noncoherent.c | 47 void arch_dma_prep_coherent(struct page *page, size_t size) argument 49 dma_cache_wback_inv((unsigned long)page_address(page), size); 98 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); local 106 if (PageHighMem(page)) { 111 addr = kmap_atomic(page); 119 page++;
|
/linux-master/fs/f2fs/ |
H A D | node.h | 239 static inline nid_t ino_of_node(struct page *node_page) 245 static inline nid_t nid_of_node(struct page *node_page) 251 static inline unsigned int ofs_of_node(struct page *node_page) 258 static inline __u64 cpver_of_node(struct page *node_page) 264 static inline block_t next_blkaddr_of_node(struct page *node_page) 270 static inline void fill_node_footer(struct page *page, nid_t nid, argument 273 struct f2fs_node *rn = F2FS_NODE(page); 289 static inline void copy_node_footer(struct page *dst, struct page *sr 296 fill_node_footer_blkaddr(struct page *page, block_t blkaddr) argument 309 is_recoverable_dnode(struct page *page) argument 392 is_node(struct page *page, int type) argument 402 set_cold_node(struct page *page, bool is_dir) argument 414 set_mark(struct page *page, int mark, int type) argument 425 f2fs_inode_chksum_set(F2FS_P_SB(page), page); local [all...] |
/linux-master/tools/testing/radix-tree/ |
H A D | regression2.c | 63 struct page { struct 67 static struct page *page_alloc(void) 69 struct page *p; 70 p = malloc(sizeof(struct page)); 79 struct page *p; 82 struct page *pages[1];
|
/linux-master/fs/hfsplus/ |
H A D | bitmap.c | 23 struct page *page; local 37 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, NULL); 38 if (IS_ERR(page)) { 42 pptr = kmap_local_page(page); 81 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, 83 if (IS_ERR(page)) { 87 curr = pptr = kmap_local_page(page); 129 set_page_dirty(page); 132 page 167 struct page *page; local [all...] |
/linux-master/arch/xtensa/mm/ |
H A D | cache.c | 33 #include <asm/page.h> 37 * The kernel provides one architecture bit PG_arch_1 in the page flags that 44 * are coherent. The kernel clears this bit whenever a page is added to the 45 * page cache. At that time, the caches might not be in sync. We, therefore, 52 * page. 59 static inline void kmap_invalidate_coherent(struct page *page, argument 62 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { 65 if (!PageHighMem(page)) { 66 kvaddr = (unsigned long)page_to_virt(page); 81 coherent_kvaddr(struct page *page, unsigned long base, unsigned long vaddr, unsigned long *paddr) argument 88 clear_user_highpage(struct page *page, unsigned long vaddr) argument 271 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 311 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument [all...] |