/linux-master/include/linux/ |
H A D | mmdebug.h | 8 struct page; 13 void dump_page(const struct page *page, const char *reason); 20 #define VM_BUG_ON_PAGE(cond, page) \ 23 dump_page(page, "VM_BUG_ON_PAGE(" __stringify(cond)")");\ 30 dump_page(&folio->page, "VM_BUG_ON_FOLIO(" __stringify(cond)")");\ 48 #define VM_WARN_ON_ONCE_PAGE(cond, page) ({ \ 53 dump_page(page, "VM_WARN_ON_ONCE_PAGE(" __stringify(cond)")");\ 63 dump_page(&folio->page, "VM_WARN_ON_FOLIO(" __stringify(cond)")");\ 73 dump_page(&folio->page, "VM_WARN_ON_ONCE_FOLI [all...] |
H A D | kasan.h | 13 struct page; 109 void __kasan_poison_pages(struct page *page, unsigned int order, bool init); 110 static __always_inline void kasan_poison_pages(struct page *page, argument 114 __kasan_poison_pages(page, order, init); 117 bool __kasan_unpoison_pages(struct page *page, unsigned int order, bool init); 118 static __always_inline bool kasan_unpoison_pages(struct page *page, argument 253 kasan_mempool_poison_pages(struct page *page, unsigned int order) argument 275 kasan_mempool_unpoison_pages(struct page *page, unsigned int order) argument 357 kasan_poison_pages(struct page *page, unsigned int order, bool init) argument 359 kasan_unpoison_pages(struct page *page, unsigned int order, bool init) argument 398 kasan_mempool_poison_pages(struct page *page, unsigned int order) argument 402 kasan_mempool_unpoison_pages(struct page *page, unsigned int order) argument [all...] |
/linux-master/mm/ |
H A D | gup.c | 33 static inline void sanity_check_pinned_pages(struct page **pages, 42 * stick around until the page is freed. 47 * THP we can assume that either the given page (PTE-mapped THP) or 48 * the head page (PMD-mapped THP) should be PageAnonExclusive(). If 52 struct page *page = *pages; local 53 struct folio *folio = page_folio(page); 55 if (is_zero_page(page) || 59 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page), page); 71 try_get_folio(struct page *page, int refs) argument 126 try_grab_folio(struct page *page, int refs, unsigned int flags) argument 225 try_grab_page(struct page *page, unsigned int flags) argument 272 unpin_user_page(struct page *page) argument 424 unpin_user_page_range_dirty_lock(struct page *page, unsigned long npages, bool make_dirty) argument 505 record_subpages(struct page *page, unsigned long sz, unsigned long addr, unsigned long end, struct page **pages) argument 541 struct page *page; local 614 struct page *page; local 691 struct page *page; local 742 can_follow_write_pmd(pmd_t pmd, struct page *page, struct vm_area_struct *vma, unsigned int flags) argument 786 struct page *page; local 863 can_follow_write_pte(pte_t pte, struct page *page, struct vm_area_struct *vma, unsigned int flags) argument 905 struct page *page; local 1019 struct page *page; local 1074 struct page *page; local 1149 struct page *page; local 1173 struct page *page; local 1191 get_gate_page(struct mm_struct *mm, unsigned long address, unsigned int gup_flags, struct vm_area_struct **vma, struct page **page) argument 1535 struct page *page; local 2371 struct page *page; local 2934 struct page *page; local 3039 struct page *page = pfn_to_page(pfn); local 3124 struct page *page; local 3168 struct page *page; local 3214 struct page *page; local [all...] |
H A D | zsmalloc.c | 16 * struct page(s) to form a zspage. 18 * Usage of struct page fields: 19 * page->private: points to zspage 20 * page->index: links together all component pages of a zspage 21 * For the huge page, this is always 0, so we use this field 23 * page->page_type: first object offset in a subpage of zspage 25 * Usage of struct page flags: 26 * PG_private: identifies the first component page 27 * PG_owner_priv_1: identifies the huge component page 71 * span more than 1 page whic 428 is_first_page(struct page *page) argument 453 get_first_obj_offset(struct page *page) argument 458 set_first_obj_offset(struct page *page, unsigned int offset) argument 713 get_zspage(struct page *page) argument 721 get_next_page(struct page *page) argument 737 obj_to_location(unsigned long obj, struct page **page, unsigned int *obj_idx) argument 744 obj_to_page(unsigned long obj, struct page **page) argument 754 location_to_obj(struct page *page, unsigned int obj_idx) argument 769 obj_allocated(struct page *page, void *obj, unsigned long *phandle) argument 789 reset_page(struct page *page) argument 822 struct page *page, *next; local 871 struct page *page = get_first_page(zspage); local 915 struct page *page; local 962 struct page *page; local 1164 struct page *page; local 1225 struct page *page; local 1522 find_alloced_obj(struct size_class *class, struct page *page, int *obj_idx) argument 1642 struct page *curr_page, *page; local 1711 struct page *page; local 1731 zs_page_isolate(struct page *page, isolate_mode_t mode) argument 1737 VM_BUG_ON_PAGE(PageIsolated(page), page); local 1742 zs_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) argument 1822 zs_page_putback(struct page *page) argument 1885 struct page *page = get_first_page(zspage); local [all...] |
H A D | hugetlb_vmemmap.c | 23 * struct vmemmap_remap_walk - walk vmemmap page table 27 * @reuse_page: the page which is reused for the tail vmemmap pages. 28 * @reuse_addr: the virtual address of the @reuse_page page. 31 * @flags: used to modify behavior in vmemmap page table walking 38 struct page *reuse_page; 49 static int vmemmap_split_pmd(pmd_t *pmd, struct page *head, unsigned long start, 99 struct page *head; 112 * the vmemmap page associated with the first vmemmap page 123 struct page *pag local 185 free_vmemmap_page(struct page *page) argument 196 struct page *page, *next; local 210 struct page *page = pte_page(ptep_get(pte)); local 254 struct page *page; local 386 struct page *page, *next; local [all...] |
H A D | memory-failure.c | 13 * Handles page cache pages in various states. The tricky part 14 * here is that we can access any page asynchronously in respect to 26 * - The case actually shows up as a frequent (top 10) page state in 27 * tools/mm/page-types when running a real workload. 41 #include <linux/page-flags.h> 148 * 1: the page is dissolved (if needed) and taken off from buddy, 149 * 0: the page is dissolved (if needed) and not taken off from buddy, 152 static int __page_handle_poison(struct page *page) argument 162 * Disabling pcp before dissolving the page wa 177 page_handle_poison(struct page *page, bool hugepage_or_freepage, bool release) argument 388 shake_page(struct page *page) argument 608 collect_procs_anon(struct folio *folio, struct page *page, struct list_head *to_kill, int force_early) argument 645 collect_procs_file(struct folio *folio, struct page *page, struct list_head *to_kill, int force_early) argument 693 collect_procs_fsdax(struct page *page, struct address_space *mapping, pgoff_t pgoff, struct list_head *to_kill, bool pre_remove) argument 727 collect_procs(struct folio *folio, struct page *page, struct list_head *tokill, int force_early) argument 1362 PageHWPoisonTakenOff(struct page *page) argument 1367 SetPageHWPoisonTakenOff(struct page *page) argument 1372 ClearPageHWPoisonTakenOff(struct page *page) argument 1384 HWPoisonHandlable(struct page *page, unsigned long flags) argument 1396 __get_hwpoison_page(struct page *page, unsigned long flags) argument 1493 __get_unpoison_page(struct page *page) argument 1691 try_to_split_thp_page(struct page *page) argument 1810 struct page *page; local 1850 struct page *page; member in struct:raw_hwp_page 1858 is_raw_hwpoison_page_in_hugepage(struct page *page) argument 1911 folio_set_hugetlb_hwpoison(struct folio *folio, struct page *page) argument 1996 struct page *page = pfn_to_page(pfn); local 2128 struct page *page; local 2673 soft_offline_in_use_page(struct page *page) argument 2772 struct page *page; local [all...] |
H A D | z3fold.c | 11 * can store up to three compressed pages per page which improves the 13 * storing an integral number of objects per page) and simplicity. 16 * number of object per page) when reclaim is used. 32 #include <linux/page-flags.h> 49 * in the beginning of an allocated page are occupied by z3fold header, so 51 * which shows the max number of free chunks in z3fold page, also there will 92 * struct z3fold_header - z3fold page metadata occupying first chunks of each 93 * z3fold page, except for HEADLESS pages 94 * @buddy: links the z3fold page into the relevant list in the 96 * @page_lock: per-page loc 256 struct page *page = virt_to_page(zhdr); local 273 struct page *page = virt_to_page(zhdr); local 319 init_z3fold_page(struct page *page, bool headless, struct z3fold_pool *pool, gfp_t gfp) argument 350 free_z3fold_page(struct page *page, bool headless) argument 440 struct page *page = virt_to_page(zhdr); local 500 struct page *page = virt_to_page(zhdr); local 690 struct page *page = virt_to_page(zhdr); local 738 struct page *page; local 790 struct page *page; local 1006 struct page *page = NULL; local 1096 struct page *page; local 1178 struct page *page; local 1223 struct page *page; local 1250 z3fold_page_isolate(struct page *page, isolate_mode_t mode) argument 1255 VM_BUG_ON_PAGE(PageIsolated(page), page); local 1286 z3fold_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) argument 1345 z3fold_page_putback(struct page *page) argument [all...] |
/linux-master/drivers/target/iscsi/ |
H A D | iscsi_target_stat.c | 51 char *page) 53 return snprintf(page, PAGE_SIZE, "%u\n", 58 char *page) 60 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); 64 char *page) 66 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); 70 char *page) 72 return snprintf(page, PAGE_SIZE, "%u\n", 77 char *page) 79 return snprintf(page, PAGE_SIZ 50 iscsi_stat_instance_inst_show(struct config_item *item, char *page) argument 57 iscsi_stat_instance_min_ver_show(struct config_item *item, char *page) argument 63 iscsi_stat_instance_max_ver_show(struct config_item *item, char *page) argument 69 iscsi_stat_instance_portals_show(struct config_item *item, char *page) argument 76 iscsi_stat_instance_nodes_show(struct config_item *item, char *page) argument 82 iscsi_stat_instance_sessions_show(struct config_item *item, char *page) argument 89 iscsi_stat_instance_fail_sess_show(struct config_item *item, char *page) argument 105 iscsi_stat_instance_fail_type_show(struct config_item *item, char *page) argument 115 iscsi_stat_instance_fail_rem_name_show(struct config_item *item, char *page) argument 126 iscsi_stat_instance_disc_time_show(struct config_item *item, char *page) argument 132 iscsi_stat_instance_description_show(struct config_item *item, char *page) argument 138 iscsi_stat_instance_vendor_show(struct config_item *item, char *page) argument 144 iscsi_stat_instance_version_show(struct config_item *item, char *page) argument 196 iscsi_stat_sess_err_inst_show(struct config_item *item, char *page) argument 203 iscsi_stat_sess_err_digest_errors_show(struct config_item *item, char *page) argument 212 iscsi_stat_sess_err_cxn_errors_show(struct config_item *item, char *page) argument 221 iscsi_stat_sess_err_format_errors_show(struct config_item *item, char *page) argument 258 iscsi_stat_tgt_attr_inst_show(struct config_item *item, char *page) argument 265 iscsi_stat_tgt_attr_indx_show(struct config_item *item, char *page) argument 271 iscsi_stat_tgt_attr_login_fails_show(struct config_item *item, char *page) argument 287 iscsi_stat_tgt_attr_last_fail_time_show(struct config_item *item, char *page) argument 303 iscsi_stat_tgt_attr_last_fail_type_show(struct config_item *item, char *page) argument 317 iscsi_stat_tgt_attr_fail_intr_name_show(struct config_item *item, char *page) argument 332 iscsi_stat_tgt_attr_fail_intr_addr_type_show(struct config_item *item, char *page) argument 349 iscsi_stat_tgt_attr_fail_intr_addr_show(struct config_item *item, char *page) argument 399 iscsi_stat_login_inst_show(struct config_item *item, char *page) argument 405 iscsi_stat_login_indx_show(struct config_item *item, char *page) argument 411 iscsi_stat_login_accepts_show(struct config_item *item, char *page) argument 425 iscsi_stat_login_other_fails_show(struct config_item *item, char *page) argument 439 iscsi_stat_login_redirects_show(struct config_item *item, char *page) argument 453 iscsi_stat_login_authorize_fails_show(struct config_item *item, char *page) argument 467 iscsi_stat_login_authenticate_fails_show( struct config_item *item, char *page) argument 481 iscsi_stat_login_negotiate_fails_show(struct config_item *item, char *page) argument 531 iscsi_stat_logout_inst_show(struct config_item *item, char *page) argument 537 iscsi_stat_logout_indx_show(struct config_item *item, char *page) argument 542 iscsi_stat_logout_normal_logouts_show(struct config_item *item, char *page) argument 551 iscsi_stat_logout_abnormal_logouts_show(struct config_item *item, char *page) argument 588 iscsi_stat_sess_inst_show(struct config_item *item, char *page) argument 598 iscsi_stat_sess_node_show(struct config_item *item, char *page) argument 619 iscsi_stat_sess_indx_show(struct config_item *item, char *page) argument 640 iscsi_stat_sess_cmd_pdus_show(struct config_item *item, char *page) argument 662 iscsi_stat_sess_rsp_pdus_show(struct config_item *item, char *page) argument 684 iscsi_stat_sess_txdata_octs_show(struct config_item *item, char *page) argument 706 iscsi_stat_sess_rxdata_octs_show(struct config_item *item, char *page) argument 728 iscsi_stat_sess_conn_digest_errors_show(struct config_item *item, char *page) argument 750 iscsi_stat_sess_conn_timeout_errors_show( struct config_item *item, char *page) argument [all...] |
H A D | iscsi_target_configfs.c | 39 static ssize_t lio_target_np_driver_show(struct config_item *item, char *page, argument 48 rb = sysfs_emit(page, "1\n"); 50 rb = sysfs_emit(page, "0\n"); 56 const char *page, size_t count, enum iscsit_transport_type type, 66 rc = kstrtou32(page, 0, &op); 116 static ssize_t lio_target_np_iser_show(struct config_item *item, char *page) argument 118 return lio_target_np_driver_show(item, page, ISCSI_INFINIBAND); 122 const char *page, size_t count) 124 return lio_target_np_driver_store(item, page, count, 129 static ssize_t lio_target_np_cxgbit_show(struct config_item *item, char *page) argument 55 lio_target_np_driver_store(struct config_item *item, const char *page, size_t count, enum iscsit_transport_type type, const char *mod_name) argument 121 lio_target_np_iser_store(struct config_item *item, const char *page, size_t count) argument 134 lio_target_np_cxgbit_store(struct config_item *item, const char *page, size_t count) argument 317 iscsi_nacl_attrib_authentication_show(struct config_item *item, char *page) argument 326 iscsi_nacl_attrib_authentication_store(struct config_item *item, const char *page, size_t count) argument 524 lio_target_nacl_info_show(struct config_item *item, char *page) argument 641 lio_target_nacl_cmdsn_depth_show(struct config_item *item, char *page) argument 647 lio_target_nacl_cmdsn_depth_store(struct config_item *item, const char *page, size_t count) argument 696 lio_target_nacl_tag_show(struct config_item *item, char *page) argument 701 lio_target_nacl_tag_store(struct config_item *item, const char *page, size_t count) argument 1023 lio_target_tpg_dynamic_sessions_show(struct config_item *item, char *page) argument 1136 lio_target_wwn_lio_version_show(struct config_item *item, char *page) argument 1144 lio_target_wwn_cpus_allowed_list_show( struct config_item *item, char *page) argument 1151 lio_target_wwn_cpus_allowed_list_store( struct config_item *item, const char *page, size_t count) argument 1279 iscsi_disc_enforce_discovery_auth_show(struct config_item *item, char *page) argument 1287 iscsi_disc_enforce_discovery_auth_store(struct config_item *item, const char *page, size_t count) argument [all...] |
/linux-master/drivers/most/ |
H A D | configfs.c | 114 const char *page, size_t count) 120 ret = kstrtobool(page, &tmp); 136 const char *page, size_t count) 142 ret = kstrtobool(page, &tmp); 160 static ssize_t mdev_link_direction_show(struct config_item *item, char *page) argument 162 return snprintf(page, PAGE_SIZE, "%s\n", to_mdev_link(item)->direction); 166 const char *page, size_t count) 170 if (!sysfs_streq(page, "dir_rx") && !sysfs_streq(page, "rx") && 171 !sysfs_streq(page, "dir_t 113 mdev_link_create_link_store(struct config_item *item, const char *page, size_t count) argument 135 mdev_link_destroy_link_store(struct config_item *item, const char *page, size_t count) argument 165 mdev_link_direction_store(struct config_item *item, const char *page, size_t count) argument 178 mdev_link_datatype_show(struct config_item *item, char *page) argument 183 mdev_link_datatype_store(struct config_item *item, const char *page, size_t count) argument 197 mdev_link_device_show(struct config_item *item, char *page) argument 202 mdev_link_device_store(struct config_item *item, const char *page, size_t count) argument 212 mdev_link_channel_show(struct config_item *item, char *page) argument 217 mdev_link_channel_store(struct config_item *item, const char *page, size_t count) argument 227 mdev_link_comp_show(struct config_item *item, char *page) argument 232 mdev_link_comp_store(struct config_item *item, const char *page, size_t count) argument 242 mdev_link_comp_params_show(struct config_item *item, char *page) argument 248 mdev_link_comp_params_store(struct config_item *item, const char *page, size_t count) argument 258 mdev_link_num_buffers_show(struct config_item *item, char *page) argument 264 mdev_link_num_buffers_store(struct config_item *item, const char *page, size_t count) argument 276 mdev_link_buffer_size_show(struct config_item *item, char *page) argument 282 mdev_link_buffer_size_store(struct config_item *item, const char *page, size_t count) argument 294 mdev_link_subbuffer_size_show(struct config_item *item, char *page) argument 301 mdev_link_subbuffer_size_store(struct config_item *item, const char *page, size_t count) argument 313 mdev_link_packets_per_xact_show(struct config_item *item, char *page) argument 320 mdev_link_packets_per_xact_store(struct config_item *item, const char *page, size_t count) argument 332 mdev_link_dbr_size_show(struct config_item *item, char *page) argument 337 mdev_link_dbr_size_store(struct config_item *item, const char *page, size_t count) argument 540 most_snd_grp_create_card_store(struct config_item *item, const char *page, size_t count) argument [all...] |
/linux-master/fs/vboxsf/ |
H A D | file.c | 9 #include <linux/page-flags.h> 189 * last known mtime and invalidate the page-cache if they differ. 231 struct page *page = &folio->page; local 233 loff_t off = page_offset(page); 238 buf = kmap(page); 243 flush_dcache_page(page); 244 SetPageUptodate(page); 246 SetPageError(page); 272 vboxsf_writepage(struct page *page, struct writeback_control *wbc) argument 309 vboxsf_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned int len, unsigned int copied, struct page *page, void *fsdata) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | dma-noncoherent.c | 52 * In this case, each page of a buffer must be kmapped/kunmapped 57 * beyond the first page. 59 static inline void __dma_sync_page_highmem(struct page *page, argument 71 start = (unsigned long)kmap_atomic(page + seg_nr) + seg_offset; 92 * takes a struct page instead of a virtual address 96 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); local 100 __dma_sync_page_highmem(page, offset, size, dir); 102 unsigned long start = (unsigned long)page_address(page) 119 arch_dma_prep_coherent(struct page *page, size_t size) argument [all...] |
/linux-master/arch/arm/mm/ |
H A D | flush.c | 128 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, argument 140 flush_pfn_alias(page_to_pfn(page), uaddr); 149 flush_icache_alias(page_to_pfn(page), uaddr, len); 159 void flush_ptrace_access(struct vm_area_struct *vma, struct page *page, argument 167 __flush_ptrace_access(page, uaddr, kaddr, len, flags); 170 void flush_uprobe_xol_access(struct page *page, unsigned long uaddr, argument 175 __flush_ptrace_access(page, uadd 185 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long uaddr, void *dst, const void *src, unsigned long len) argument 367 flush_dcache_page(struct page *page) argument 382 __flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument [all...] |
H A D | dma-mapping.c | 28 #include <asm/page.h> 55 struct page *page; member in struct:arm_dma_free_args 64 struct page **ret_page); 107 static void __dma_clear_buffer(struct page *page, size_t size, int coherent_flag) argument 113 if (PageHighMem(page)) { 114 phys_addr_t base = __pfn_to_phys(page_to_pfn(page)); 117 void *ptr = kmap_atomic(page); 122 page 145 struct page *page, *p, *e; local 166 __dma_free_buffer(struct page *page, size_t size) argument 204 struct page *page; local 312 struct page *page = virt_to_page((void *)addr); local 319 __dma_remap(struct page *page, size_t size, pgprot_t prot) argument 332 struct page *page; local 398 struct page *page; local 426 __free_from_contiguous(struct device *dev, struct page *page, void *cpu_addr, size_t size, bool want_vaddr) argument 449 struct page *page; local 538 struct page *page = NULL; local 609 struct page *page = phys_to_page(dma_to_phys(dev, handle)); local 627 dma_cache_maint_page(struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, void (*op)(const void *, size_t, int)) argument 679 __dma_page_cpu_to_dev(struct page *page, unsigned long off, size_t size, enum dma_data_direction dir) argument 695 __dma_page_dev_to_cpu(struct page *page, unsigned long off, size_t size, enum dma_data_direction dir) argument 869 struct page *page; local 1020 struct page *page; local 1044 struct page *page; local 1368 arm_iommu_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, unsigned long attrs) argument 1410 struct page *page; local 1488 struct page *page; local 1503 struct page *page; local [all...] |
/linux-master/drivers/block/ |
H A D | brd.c | 47 * Look up and return a brd's page for a given sector. 49 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) 55 * Insert a new page for a given sector, if one does not already exist. 60 struct page *page; local 63 page = brd_lookup_page(brd, sector); 64 if (page) 67 page = alloc_page(gfp | __GFP_ZERO | __GFP_HIGHMEM); 68 if (!page) 72 ret = __xa_insert(&brd->brd_pages, idx, page, gf 91 struct page *page; local 129 struct page *page; local 161 struct page *page; local 192 brd_do_bvec(struct brd_device *brd, struct page *page, unsigned int len, unsigned int off, blk_opf_t opf, sector_t sector) argument 228 struct page *page; local [all...] |
/linux-master/fs/ocfs2/ |
H A D | mmap.c | 42 vma, vmf->page, vmf->pgoff); 47 struct buffer_head *di_bh, struct page *page) 53 loff_t pos = page_offset(page); 56 struct page *locked_page = NULL; 63 * There are cases that lead to the page no longer belonging to the 70 * then re-extended the file. We'll re-check the page mapping after 71 * taking the page lock inside of ocfs2_write_begin_nolock(). 75 if ((page->mapping != inode->i_mapping) || 76 (!PageUptodate(page)) || 46 __ocfs2_page_mkwrite(struct file *file, struct buffer_head *di_bh, struct page *page) argument 115 struct page *page = vmf->page; local [all...] |
/linux-master/arch/x86/kernel/ |
H A D | espfix_64.c | 22 * This file sets up the ministacks and the related page tables. The 61 /* Page allocation bitmap - each page serves ESPFIX_STACKS_PER_PAGE CPUs */ 73 * we have to account for some amount of padding at the end of each page. 77 unsigned long page, slot; local 80 page = (cpu / ESPFIX_STACKS_PER_PAGE) ^ page_random; 82 addr = (page << PAGE_SHIFT) + (slot * ESPFIX_STACK_SIZE); 113 /* Install the espfix pud into the kernel page directory */ 127 unsigned int page; local 145 page = cpu/ESPFIX_STACKS_PER_PAGE; 148 stack_page = READ_ONCE(espfix_pages[page]); 165 struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0); local 177 struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0); local [all...] |
/linux-master/drivers/hwmon/pmbus/ |
H A D | stpddc60.c | 85 static int stpddc60_read_byte_data(struct i2c_client *client, int page, int reg) argument 89 if (page > 0) 109 static int stpddc60_read_word_data(struct i2c_client *client, int page, argument 114 if (page > 0) 119 ret = pmbus_read_word_data(client, page, phase, 127 ret = pmbus_read_word_data(client, page, phase, reg); 145 static int stpddc60_write_word_data(struct i2c_client *client, int page, argument 151 if (page > 0) 156 ret = pmbus_read_word_data(client, page, 0xff, 161 ret = pmbus_write_byte_data(client, page, [all...] |
H A D | ltc3815.c | 27 static int ltc3815_read_byte_data(struct i2c_client *client, int page, int reg) argument 48 static int ltc3815_write_byte(struct i2c_client *client, int page, u8 reg) argument 72 static int ltc3815_read_word_data(struct i2c_client *client, int page, argument 79 ret = pmbus_read_word_data(client, page, phase, 83 ret = pmbus_read_word_data(client, page, phase, 87 ret = pmbus_read_word_data(client, page, phase, 91 ret = pmbus_read_word_data(client, page, phase, 95 ret = pmbus_read_word_data(client, page, phase, 112 static int ltc3815_write_word_data(struct i2c_client *client, int page, argument 119 ret = pmbus_write_word_data(client, page, [all...] |
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | main.c | 51 * from the input list, and made available for the page allocator. SECS pages 60 struct sgx_epc_page *page; local 69 page = list_first_entry(dirty_page_list, struct sgx_epc_page, list); 72 * Checking page->poison without holding the node->lock 75 * called for a page that sgx_free_epc_page() will put onto 78 if (page->poison) { 79 struct sgx_epc_section *section = &sgx_epc_sections[page->section]; 83 list_move(&page->list, &node->sgx_poison_page_list); 89 ret = __eremove(sgx_get_epc_virt_addr(page)); 92 * page i 112 struct sgx_encl_page *page = epc_page->owner; local 144 struct sgx_encl_page *page = epc_page->owner; local 445 struct sgx_epc_page *page = NULL; local 476 struct sgx_epc_page *page; local 507 sgx_mark_page_reclaimable(struct sgx_epc_page *page) argument 525 sgx_unmark_page_reclaimable(struct sgx_epc_page *page) argument 562 struct sgx_epc_page *page; local 603 sgx_free_epc_page(struct sgx_epc_page *page) argument 679 struct sgx_epc_page *page = sgx_paddr_to_page(pfn << PAGE_SHIFT); local [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, 54 * have to bother with retries or mapping a dummy page to insure we 59 int qib_map_page(struct pci_dev *hwdev, struct page *page, dma_addr_t *daddr) argument 63 phys = dma_map_page(&hwdev->dev, page, 0, PAGE_SIZE, DMA_FROM_DEVICE); 69 phys = dma_map_page(&hwdev->dev, page, 0, PAGE_SIZE, 74 * FIXME: If we get 0 again, we should keep this page, 75 * map another, then free the 0 page. 84 * @start_page: the start page 86 * @p: the output page structure [all...] |
/linux-master/arch/arc/include/asm/ |
H A D | page.h | 8 #include <uapi/asm/page.h> 29 struct page; 33 void copy_user_highpage(struct page *to, struct page *from, 35 void clear_user_page(void *to, unsigned long u_vaddr, struct page *page); 85 typedef struct page *pgtable_t;
|
/linux-master/arch/csky/include/asm/ |
H A D | page.h | 11 * PAGE_SHIFT determines the page size: 4KB 43 #define clear_page(page) memset((page), 0, PAGE_SIZE) 46 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) 49 struct page; 51 #include <abi/page.h> 60 typedef struct page *pgtable_t;
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_db.c | 14 struct hns_roce_user_db_page *page; local 20 list_for_each_entry(page, &context->page_list, list) 21 if (page->user_virt == page_addr) 24 page = kmalloc(sizeof(*page), GFP_KERNEL); 25 if (!page) { 30 refcount_set(&page->refcount, 1); 31 page->user_virt = page_addr; 32 page->umem = ib_umem_get(context->ibucontext.device, page_addr, 34 if (IS_ERR(page [all...] |
/linux-master/kernel/dma/ |
H A D | pool.c | 83 struct page *page = NULL; local 93 page = dma_alloc_from_contiguous(NULL, 1 << order, 95 if (!page) 96 page = alloc_pages(gfp, order); 97 } while (!page && order-- > 0); 98 if (!page) 101 arch_dma_prep_coherent(page, pool_size); 104 addr = dma_common_contiguous_remap(page, pool_size, 110 addr = page_to_virt(page); 270 struct page *page; local [all...] |