/linux-master/arch/csky/include/asm/ |
H A D | tlbflush.h | 11 * - flush_tlb_page(vma, vmaddr) flushes one page 17 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
|
/linux-master/arch/s390/kernel/vdso32/ |
H A D | vdso32_wrapper.S | 4 #include <asm/page.h>
|
/linux-master/drivers/dma/ioat/ |
H A D | sysfs.c | 17 static ssize_t cap_show(struct dma_chan *c, char *page) argument 21 return sprintf(page, "copy%s%s%s%s%s\n", 31 static ssize_t version_show(struct dma_chan *c, char *page) argument 36 return sprintf(page, "%d.%d\n", 42 ioat_attr_show(struct kobject *kobj, struct attribute *attr, char *page) argument 52 return entry->show(&ioat_chan->dma_chan, page); 57 const char *page, size_t count) 67 return entry->store(&ioat_chan->dma_chan, page, count); 111 static ssize_t ring_size_show(struct dma_chan *c, char *page) argument 115 return sprintf(page, " 56 ioat_attr_store(struct kobject *kobj, struct attribute *attr, const char *page, size_t count) argument 119 ring_active_show(struct dma_chan *c, char *page) argument 128 intr_coalesce_show(struct dma_chan *c, char *page) argument 135 intr_coalesce_store(struct dma_chan *c, const char *page, size_t count) argument [all...] |
/linux-master/arch/parisc/kernel/vdso32/ |
H A D | vdso32_wrapper.S | 3 #include <asm/page.h>
|
/linux-master/arch/parisc/kernel/vdso64/ |
H A D | vdso64_wrapper.S | 3 #include <asm/page.h>
|
/linux-master/arch/powerpc/kernel/ |
H A D | vdso32_wrapper.S | 3 #include <asm/page.h>
|
H A D | vdso64_wrapper.S | 3 #include <asm/page.h>
|
/linux-master/arch/s390/include/asm/ |
H A D | user.h | 11 #include <asm/page.h> 22 UPAGE: 1 page consisting of a user struct that tells gdb what is present 26 always be only one page. 29 that may have been malloced. No attempt is made to determine if a page 30 is demand-zero or if a page is totally unused, we just cover the entire
|
/linux-master/arch/loongarch/vdso/ |
H A D | vdso.S | 11 #include <asm/page.h>
|
/linux-master/include/linux/ |
H A D | kernel-page-flags.h | 5 #include <uapi/linux/kernel-page-flags.h>
|
H A D | page_ext.h | 46 * A page_ext page is associated with every page descriptor. The 47 * page_ext helps us add more information about the page. 80 extern struct page_ext *page_ext_get(const struct page *page); 120 static inline struct page_ext *page_ext_get(const struct page *page) argument
|
/linux-master/arch/x86/entry/vdso/vdso32/ |
H A D | vdso32.lds.S | 10 #include <asm/page.h>
|
/linux-master/arch/xtensa/include/asm/ |
H A D | tlb.h | 15 #include <asm/page.h>
|
/linux-master/arch/arm/include/asm/ |
H A D | kexec.h | 69 static inline unsigned long page_to_boot_pfn(struct page *page) argument 71 return page_to_pfn(page) + (arch_phys_to_idmap_offset >> PAGE_SHIFT); 75 static inline struct page *boot_pfn_to_page(unsigned long boot_pfn)
|
/linux-master/fs/jfs/ |
H A D | jfs_metapage.h | 23 sector_t index; /* block address of page */ 27 struct page *page; member in struct:metapage 65 * hold_metapage and put_metapage are used in conjunction. The page lock 93 struct page *page = mp->page; local 94 lock_page(page); 97 get_page(page); 98 wait_on_page_writeback(page); [all...] |
/linux-master/include/xen/interface/io/ |
H A D | kbdif.h | 146 * page-gref 150 * a sole page in a single page sized event ring buffer. 152 * page-ref 156 * PFN of the shared page. 192 #define XENKBD_FIELD_RING_GREF "page-gref" 202 #define XENKBD_FIELD_RING_REF "page-ref" 210 * each other using a shared page and an event channel. 211 * Shared page contains a ring with event structures. 518 * Shared page [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | page_32.h | 20 static inline void clear_page(void *page) argument 22 memset(page, 0, PAGE_SIZE);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io-buffered.h | 14 unsigned, struct page **, void **); 16 unsigned, unsigned, struct page *, void *);
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_fw_mips.h | 9 #include <asm/page.h> 25 struct page *pt_pages[PVR_MIPS_PT_PAGE_COUNT];
|
/linux-master/arch/mips/mm/ |
H A D | dma-noncoherent.c | 47 void arch_dma_prep_coherent(struct page *page, size_t size) argument 49 dma_cache_wback_inv((unsigned long)page_address(page), size); 98 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); local 106 if (PageHighMem(page)) { 111 addr = kmap_atomic(page); 119 page++;
|
/linux-master/fs/f2fs/ |
H A D | node.h | 239 static inline nid_t ino_of_node(struct page *node_page) 245 static inline nid_t nid_of_node(struct page *node_page) 251 static inline unsigned int ofs_of_node(struct page *node_page) 258 static inline __u64 cpver_of_node(struct page *node_page) 264 static inline block_t next_blkaddr_of_node(struct page *node_page) 270 static inline void fill_node_footer(struct page *page, nid_t nid, argument 273 struct f2fs_node *rn = F2FS_NODE(page); 289 static inline void copy_node_footer(struct page *dst, struct page *sr 296 fill_node_footer_blkaddr(struct page *page, block_t blkaddr) argument 309 is_recoverable_dnode(struct page *page) argument 392 is_node(struct page *page, int type) argument 402 set_cold_node(struct page *page, bool is_dir) argument 414 set_mark(struct page *page, int mark, int type) argument 425 f2fs_inode_chksum_set(F2FS_P_SB(page), page); local [all...] |
/linux-master/tools/testing/radix-tree/ |
H A D | regression2.c | 63 struct page { struct 67 static struct page *page_alloc(void) 69 struct page *p; 70 p = malloc(sizeof(struct page)); 79 struct page *p; 82 struct page *pages[1];
|
/linux-master/fs/hfsplus/ |
H A D | bitmap.c | 23 struct page *page; local 37 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, NULL); 38 if (IS_ERR(page)) { 42 pptr = kmap_local_page(page); 81 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, 83 if (IS_ERR(page)) { 87 curr = pptr = kmap_local_page(page); 129 set_page_dirty(page); 132 page 167 struct page *page; local [all...] |
/linux-master/arch/xtensa/mm/ |
H A D | cache.c | 33 #include <asm/page.h> 37 * The kernel provides one architecture bit PG_arch_1 in the page flags that 44 * are coherent. The kernel clears this bit whenever a page is added to the 45 * page cache. At that time, the caches might not be in sync. We, therefore, 52 * page. 59 static inline void kmap_invalidate_coherent(struct page *page, argument 62 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { 65 if (!PageHighMem(page)) { 66 kvaddr = (unsigned long)page_to_virt(page); 81 coherent_kvaddr(struct page *page, unsigned long base, unsigned long vaddr, unsigned long *paddr) argument 88 clear_user_highpage(struct page *page, unsigned long vaddr) argument 273 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 313 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument [all...] |
/linux-master/mm/ |
H A D | highmem.c | 13 * Rewrote high memory support to move the page cache into 62 * Determine color of virtual address where the page should be mapped. 64 static inline unsigned int get_pkmap_color(struct page *page) argument 71 * Get next index for mapping inside PKMAP region for page with given color. 82 * Determine if page index inside PKMAP region (pkmap_nr) of given color 154 struct page *__kmap_to_page(void *vaddr) 193 struct page *page; local 212 * no-one has the page mappe 234 map_new_virtual(struct page *page) argument 296 kmap_high(struct page *page) argument 326 kmap_high_get(struct page *page) argument 348 kunmap_high(struct page *page) argument 392 zero_user_segments(struct page *page, unsigned start1, unsigned end1, unsigned start2, unsigned end2) argument 500 arch_kmap_local_high_get(struct page *page) argument 564 __kmap_local_page_prot(struct page *page, pgprot_t prot) argument 716 struct page *page; member in struct:page_address_map 731 page_slot(const struct page *page) argument 742 page_address(const struct page *page) argument 775 set_page_address(struct page *page, void *virtual) argument [all...] |