/linux-master/arch/s390/include/asm/ |
H A D | page.h | 18 /* PAGE_SHIFT determines the page size */ 51 #define clear_page(page) memset((page), 0, PAGE_SIZE) 55 * bypass caches when copying a page. Especially when copying huge pages 73 #define clear_user_page(page, vaddr, pg) clear_page(page) 164 struct page; 165 void arch_free_page(struct page *page, int order); 166 void arch_alloc_page(struct page *pag [all...] |
/linux-master/drivers/target/iscsi/ |
H A D | iscsi_target_stat.c | 51 char *page) 53 return snprintf(page, PAGE_SIZE, "%u\n", 58 char *page) 60 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); 64 char *page) 66 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); 70 char *page) 72 return snprintf(page, PAGE_SIZE, "%u\n", 77 char *page) 79 return snprintf(page, PAGE_SIZ 50 iscsi_stat_instance_inst_show(struct config_item *item, char *page) argument 57 iscsi_stat_instance_min_ver_show(struct config_item *item, char *page) argument 63 iscsi_stat_instance_max_ver_show(struct config_item *item, char *page) argument 69 iscsi_stat_instance_portals_show(struct config_item *item, char *page) argument 76 iscsi_stat_instance_nodes_show(struct config_item *item, char *page) argument 82 iscsi_stat_instance_sessions_show(struct config_item *item, char *page) argument 89 iscsi_stat_instance_fail_sess_show(struct config_item *item, char *page) argument 105 iscsi_stat_instance_fail_type_show(struct config_item *item, char *page) argument 115 iscsi_stat_instance_fail_rem_name_show(struct config_item *item, char *page) argument 126 iscsi_stat_instance_disc_time_show(struct config_item *item, char *page) argument 132 iscsi_stat_instance_description_show(struct config_item *item, char *page) argument 138 iscsi_stat_instance_vendor_show(struct config_item *item, char *page) argument 144 iscsi_stat_instance_version_show(struct config_item *item, char *page) argument 196 iscsi_stat_sess_err_inst_show(struct config_item *item, char *page) argument 203 iscsi_stat_sess_err_digest_errors_show(struct config_item *item, char *page) argument 212 iscsi_stat_sess_err_cxn_errors_show(struct config_item *item, char *page) argument 221 iscsi_stat_sess_err_format_errors_show(struct config_item *item, char *page) argument 258 iscsi_stat_tgt_attr_inst_show(struct config_item *item, char *page) argument 265 iscsi_stat_tgt_attr_indx_show(struct config_item *item, char *page) argument 271 iscsi_stat_tgt_attr_login_fails_show(struct config_item *item, char *page) argument 287 iscsi_stat_tgt_attr_last_fail_time_show(struct config_item *item, char *page) argument 303 iscsi_stat_tgt_attr_last_fail_type_show(struct config_item *item, char *page) argument 317 iscsi_stat_tgt_attr_fail_intr_name_show(struct config_item *item, char *page) argument 332 iscsi_stat_tgt_attr_fail_intr_addr_type_show(struct config_item *item, char *page) argument 349 iscsi_stat_tgt_attr_fail_intr_addr_show(struct config_item *item, char *page) argument 399 iscsi_stat_login_inst_show(struct config_item *item, char *page) argument 405 iscsi_stat_login_indx_show(struct config_item *item, char *page) argument 411 iscsi_stat_login_accepts_show(struct config_item *item, char *page) argument 425 iscsi_stat_login_other_fails_show(struct config_item *item, char *page) argument 439 iscsi_stat_login_redirects_show(struct config_item *item, char *page) argument 453 iscsi_stat_login_authorize_fails_show(struct config_item *item, char *page) argument 467 iscsi_stat_login_authenticate_fails_show( struct config_item *item, char *page) argument 481 iscsi_stat_login_negotiate_fails_show(struct config_item *item, char *page) argument 531 iscsi_stat_logout_inst_show(struct config_item *item, char *page) argument 537 iscsi_stat_logout_indx_show(struct config_item *item, char *page) argument 542 iscsi_stat_logout_normal_logouts_show(struct config_item *item, char *page) argument 551 iscsi_stat_logout_abnormal_logouts_show(struct config_item *item, char *page) argument 588 iscsi_stat_sess_inst_show(struct config_item *item, char *page) argument 598 iscsi_stat_sess_node_show(struct config_item *item, char *page) argument 619 iscsi_stat_sess_indx_show(struct config_item *item, char *page) argument 640 iscsi_stat_sess_cmd_pdus_show(struct config_item *item, char *page) argument 662 iscsi_stat_sess_rsp_pdus_show(struct config_item *item, char *page) argument 684 iscsi_stat_sess_txdata_octs_show(struct config_item *item, char *page) argument 706 iscsi_stat_sess_rxdata_octs_show(struct config_item *item, char *page) argument 728 iscsi_stat_sess_conn_digest_errors_show(struct config_item *item, char *page) argument 750 iscsi_stat_sess_conn_timeout_errors_show( struct config_item *item, char *page) argument [all...] |
H A D | iscsi_target_configfs.c | 39 static ssize_t lio_target_np_driver_show(struct config_item *item, char *page, argument 48 rb = sysfs_emit(page, "1\n"); 50 rb = sysfs_emit(page, "0\n"); 56 const char *page, size_t count, enum iscsit_transport_type type, 66 rc = kstrtou32(page, 0, &op); 116 static ssize_t lio_target_np_iser_show(struct config_item *item, char *page) argument 118 return lio_target_np_driver_show(item, page, ISCSI_INFINIBAND); 122 const char *page, size_t count) 124 return lio_target_np_driver_store(item, page, count, 129 static ssize_t lio_target_np_cxgbit_show(struct config_item *item, char *page) argument 55 lio_target_np_driver_store(struct config_item *item, const char *page, size_t count, enum iscsit_transport_type type, const char *mod_name) argument 121 lio_target_np_iser_store(struct config_item *item, const char *page, size_t count) argument 134 lio_target_np_cxgbit_store(struct config_item *item, const char *page, size_t count) argument 317 iscsi_nacl_attrib_authentication_show(struct config_item *item, char *page) argument 326 iscsi_nacl_attrib_authentication_store(struct config_item *item, const char *page, size_t count) argument 524 lio_target_nacl_info_show(struct config_item *item, char *page) argument 641 lio_target_nacl_cmdsn_depth_show(struct config_item *item, char *page) argument 647 lio_target_nacl_cmdsn_depth_store(struct config_item *item, const char *page, size_t count) argument 696 lio_target_nacl_tag_show(struct config_item *item, char *page) argument 701 lio_target_nacl_tag_store(struct config_item *item, const char *page, size_t count) argument 1023 lio_target_tpg_dynamic_sessions_show(struct config_item *item, char *page) argument 1136 lio_target_wwn_lio_version_show(struct config_item *item, char *page) argument 1144 lio_target_wwn_cpus_allowed_list_show( struct config_item *item, char *page) argument 1151 lio_target_wwn_cpus_allowed_list_store( struct config_item *item, const char *page, size_t count) argument 1279 iscsi_disc_enforce_discovery_auth_show(struct config_item *item, char *page) argument 1287 iscsi_disc_enforce_discovery_auth_store(struct config_item *item, const char *page, size_t count) argument [all...] |
/linux-master/drivers/block/ |
H A D | brd.c | 32 * the pages containing the block device's contents. A brd page's ->index is 50 * Look up and return a brd's page for a given sector. 52 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) 55 struct page *page; local 57 idx = sector >> PAGE_SECTORS_SHIFT; /* sector to page index */ 58 page = xa_load(&brd->brd_pages, idx); 60 BUG_ON(page && page->index != idx); 62 return page; 71 struct page *page, *cur; local 109 struct page *page; local 147 struct page *page; local 179 struct page *page; local 210 brd_do_bvec(struct brd_device *brd, struct page *page, unsigned int len, unsigned int off, blk_opf_t opf, sector_t sector) argument [all...] |
/linux-master/drivers/most/ |
H A D | configfs.c | 114 const char *page, size_t count) 120 ret = kstrtobool(page, &tmp); 136 const char *page, size_t count) 142 ret = kstrtobool(page, &tmp); 160 static ssize_t mdev_link_direction_show(struct config_item *item, char *page) argument 162 return snprintf(page, PAGE_SIZE, "%s\n", to_mdev_link(item)->direction); 166 const char *page, size_t count) 170 if (!sysfs_streq(page, "dir_rx") && !sysfs_streq(page, "rx") && 171 !sysfs_streq(page, "dir_t 113 mdev_link_create_link_store(struct config_item *item, const char *page, size_t count) argument 135 mdev_link_destroy_link_store(struct config_item *item, const char *page, size_t count) argument 165 mdev_link_direction_store(struct config_item *item, const char *page, size_t count) argument 178 mdev_link_datatype_show(struct config_item *item, char *page) argument 183 mdev_link_datatype_store(struct config_item *item, const char *page, size_t count) argument 197 mdev_link_device_show(struct config_item *item, char *page) argument 202 mdev_link_device_store(struct config_item *item, const char *page, size_t count) argument 212 mdev_link_channel_show(struct config_item *item, char *page) argument 217 mdev_link_channel_store(struct config_item *item, const char *page, size_t count) argument 227 mdev_link_comp_show(struct config_item *item, char *page) argument 232 mdev_link_comp_store(struct config_item *item, const char *page, size_t count) argument 242 mdev_link_comp_params_show(struct config_item *item, char *page) argument 248 mdev_link_comp_params_store(struct config_item *item, const char *page, size_t count) argument 258 mdev_link_num_buffers_show(struct config_item *item, char *page) argument 264 mdev_link_num_buffers_store(struct config_item *item, const char *page, size_t count) argument 276 mdev_link_buffer_size_show(struct config_item *item, char *page) argument 282 mdev_link_buffer_size_store(struct config_item *item, const char *page, size_t count) argument 294 mdev_link_subbuffer_size_show(struct config_item *item, char *page) argument 301 mdev_link_subbuffer_size_store(struct config_item *item, const char *page, size_t count) argument 313 mdev_link_packets_per_xact_show(struct config_item *item, char *page) argument 320 mdev_link_packets_per_xact_store(struct config_item *item, const char *page, size_t count) argument 332 mdev_link_dbr_size_show(struct config_item *item, char *page) argument 337 mdev_link_dbr_size_store(struct config_item *item, const char *page, size_t count) argument 540 most_snd_grp_create_card_store(struct config_item *item, const char *page, size_t count) argument [all...] |
/linux-master/fs/vboxsf/ |
H A D | file.c | 9 #include <linux/page-flags.h> 189 * last known mtime and invalidate the page-cache if they differ. 231 struct page *page = &folio->page; local 233 loff_t off = page_offset(page); 238 buf = kmap(page); 243 flush_dcache_page(page); 244 SetPageUptodate(page); 246 SetPageError(page); 272 vboxsf_writepage(struct page *page, struct writeback_control *wbc) argument 309 vboxsf_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned int len, unsigned int copied, struct page *page, void *fsdata) argument [all...] |
/linux-master/mm/ |
H A D | zsmalloc.c | 16 * struct page(s) to form a zspage. 18 * Usage of struct page fields: 19 * page->private: points to zspage 20 * page->index: links together all component pages of a zspage 21 * For the huge page, this is always 0, so we use this field 23 * page->page_type: first object offset in a subpage of zspage 25 * Usage of struct page flags: 26 * PG_private: identifies the first component page 27 * PG_owner_priv_1: identifies the huge component page 71 * span more than 1 page whic 428 is_first_page(struct page *page) argument 453 get_first_obj_offset(struct page *page) argument 458 set_first_obj_offset(struct page *page, unsigned int offset) argument 713 get_zspage(struct page *page) argument 721 get_next_page(struct page *page) argument 737 obj_to_location(unsigned long obj, struct page **page, unsigned int *obj_idx) argument 744 obj_to_page(unsigned long obj, struct page **page) argument 754 location_to_obj(struct page *page, unsigned int obj_idx) argument 769 obj_allocated(struct page *page, void *obj, unsigned long *phandle) argument 789 reset_page(struct page *page) argument 822 struct page *page, *next; local 871 struct page *page = get_first_page(zspage); local 915 struct page *page; local 962 struct page *page; local 1164 struct page *page; local 1225 struct page *page; local 1522 find_alloced_obj(struct size_class *class, struct page *page, int *obj_idx) argument 1642 struct page *curr_page, *page; local 1711 struct page *page; local 1731 zs_page_isolate(struct page *page, isolate_mode_t mode) argument 1737 VM_BUG_ON_PAGE(PageIsolated(page), page); local 1742 zs_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) argument 1822 zs_page_putback(struct page *page) argument 1885 struct page *page = get_first_page(zspage); local [all...] |
H A D | memory-failure.c | 13 * Handles page cache pages in various states. The tricky part 14 * here is that we can access any page asynchronously in respect to 26 * - The case actually shows up as a frequent (top 10) page state in 27 * tools/mm/page-types when running a real workload. 41 #include <linux/page-flags.h> 149 * 1: the page is dissolved (if needed) and taken off from buddy, 150 * 0: the page is dissolved (if needed) and not taken off from buddy, 153 static int __page_handle_poison(struct page *page) argument 163 * Disabling pcp before dissolving the page wa 178 page_handle_poison(struct page *page, bool hugepage_or_freepage, bool release) argument 610 collect_procs_anon(struct folio *folio, struct page *page, struct list_head *to_kill, int force_early) argument 647 collect_procs_file(struct folio *folio, struct page *page, struct list_head *to_kill, int force_early) argument 691 collect_procs_fsdax(struct page *page, struct address_space *mapping, pgoff_t pgoff, struct list_head *to_kill, bool pre_remove) argument 725 collect_procs(struct folio *folio, struct page *page, struct list_head *tokill, int force_early) argument 1368 PageHWPoisonTakenOff(struct page *page) argument 1373 SetPageHWPoisonTakenOff(struct page *page) argument 1378 ClearPageHWPoisonTakenOff(struct page *page) argument 1390 HWPoisonHandlable(struct page *page, unsigned long flags) argument 1402 __get_hwpoison_page(struct page *page, unsigned long flags) argument 1499 __get_unpoison_page(struct page *page) argument 1697 try_to_split_thp_page(struct page *page) argument 1816 struct page *page; local 1856 struct page *page; member in struct:raw_hwp_page 1864 is_raw_hwpoison_page_in_hugepage(struct page *page) argument 1917 folio_set_hugetlb_hwpoison(struct folio *folio, struct page *page) argument 2002 struct page *page = pfn_to_page(pfn); local 2134 struct page *page; local 2670 soft_offline_in_use_page(struct page *page) argument 2768 struct page *page; local [all...] |
H A D | hugetlb_vmemmap.c | 23 * struct vmemmap_remap_walk - walk vmemmap page table 27 * @reuse_page: the page which is reused for the tail vmemmap pages. 28 * @reuse_addr: the virtual address of the @reuse_page page. 31 * @flags: used to modify behavior in vmemmap page table walking 38 struct page *reuse_page; 49 static int vmemmap_split_pmd(pmd_t *pmd, struct page *head, unsigned long start, 99 struct page *head; 112 * the vmemmap page associated with the first vmemmap page 123 struct page *pag local 185 free_vmemmap_page(struct page *page) argument 196 struct page *page, *next; local 210 struct page *page = pte_page(ptep_get(pte)); local 254 struct page *page; local 386 struct page *page, *next; local [all...] |
H A D | internal.h | 109 * @addr: The user virtual address the first page is mapped at. 111 * @pte: Page table entry for the first page. 124 * start_ptep must map any page of the folio. max_nr must be at least one and 125 * must be limited by the caller so scanning cannot exceed a single page table. 263 * Turn a non-refcounted page (->_refcount == 0) into refcounted with 266 static inline void set_page_refcounted(struct page *page) argument 268 VM_BUG_ON_PAGE(PageTail(page), page); local 269 VM_BUG_ON_PAGE(page_ref_count(page), pag local 364 buddy_order(struct page *page) argument 396 page_is_buddy(struct page *page, struct page *buddy, unsigned int order) argument 454 find_buddy_page_pfn(struct page *page, unsigned long pfn, unsigned int order, unsigned long *buddy_pfn) argument 512 page_rmappable_folio(struct page *page) argument 520 prep_compound_head(struct page *page, unsigned int order) argument 624 struct page *page; member in struct:capture_control 823 vma_address(struct page *page, struct vm_area_struct *vma) argument 825 VM_BUG_ON_PAGE(PageKsm(page), page); /* KSM page->index unusable */ local 1043 is_migrate_highatomic_page(struct page *page) argument 1157 gup_must_unshare(struct vm_area_struct *vma, unsigned int flags, struct page *page) argument [all...] |
H A D | z3fold.c | 11 * can store up to three compressed pages per page which improves the 13 * storing an integral number of objects per page) and simplicity. 16 * number of object per page) when reclaim is used. 32 #include <linux/page-flags.h> 49 * in the beginning of an allocated page are occupied by z3fold header, so 51 * which shows the max number of free chunks in z3fold page, also there will 92 * struct z3fold_header - z3fold page metadata occupying first chunks of each 93 * z3fold page, except for HEADLESS pages 94 * @buddy: links the z3fold page into the relevant list in the 96 * @page_lock: per-page loc 256 struct page *page = virt_to_page(zhdr); local 273 struct page *page = virt_to_page(zhdr); local 319 init_z3fold_page(struct page *page, bool headless, struct z3fold_pool *pool, gfp_t gfp) argument 350 free_z3fold_page(struct page *page, bool headless) argument 440 struct page *page = virt_to_page(zhdr); local 500 struct page *page = virt_to_page(zhdr); local 690 struct page *page = virt_to_page(zhdr); local 738 struct page *page; local 790 struct page *page; local 1006 struct page *page = NULL; local 1096 struct page *page; local 1178 struct page *page; local 1223 struct page *page; local 1250 z3fold_page_isolate(struct page *page, isolate_mode_t mode) argument 1255 VM_BUG_ON_PAGE(PageIsolated(page), page); local 1286 z3fold_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) argument 1345 z3fold_page_putback(struct page *page) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | dma-noncoherent.c | 52 * In this case, each page of a buffer must be kmapped/kunmapped 57 * beyond the first page. 59 static inline void __dma_sync_page_highmem(struct page *page, argument 71 start = (unsigned long)kmap_atomic(page + seg_nr) + seg_offset; 92 * takes a struct page instead of a virtual address 96 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); local 100 __dma_sync_page_highmem(page, offset, size, dir); 102 unsigned long start = (unsigned long)page_address(page) 119 arch_dma_prep_coherent(struct page *page, size_t size) argument [all...] |
/linux-master/include/linux/ |
H A D | pagemap.h | 145 * mapping_shrinkable - test if page cache state allows inode reclaim 146 * @mapping: the page cache mapping 171 * inodes before there is highmem pressure from the page 184 * head pointer, which allows non-resident page cache entries 315 * reduce page locking. 402 struct address_space *page_mapping(struct page *); 410 * For folios which are in the page cache, return the mapping that this 411 * page belongs to. Folios in the swap cache return the mapping of the 416 * Do not call this for folios which aren't in the page cache or swap cache. 430 * For folios which are in the page cach 446 page_file_mapping(struct page *page) argument 521 attach_page_private(struct page *page, void *data) argument 526 detach_page_private(struct page *page) argument 900 page_to_pgoff(struct page *page) argument 918 page_offset(struct page *page) argument 923 page_file_offset(struct page *page) argument 1017 trylock_page(struct page *page) argument 1062 lock_page(struct page *page) argument 1133 wait_on_page_locked(struct page *page) argument 1147 __set_page_dirty(struct page *page, struct address_space *mapping, int warn) argument 1384 struct page *page; local 1397 VM_BUG_ON_PAGE(PageTail(page), page); local 1509 page_mkwrite_check_truncate(struct page *page, struct inode *inode) argument 1547 i_blocks_per_page(struct inode *inode, struct page *page) argument [all...] |
H A D | kasan.h | 13 struct page; 109 void __kasan_poison_pages(struct page *page, unsigned int order, bool init); 110 static __always_inline void kasan_poison_pages(struct page *page, argument 114 __kasan_poison_pages(page, order, init); 117 bool __kasan_unpoison_pages(struct page *page, unsigned int order, bool init); 118 static __always_inline bool kasan_unpoison_pages(struct page *page, argument 253 kasan_mempool_poison_pages(struct page *page, unsigned int order) argument 275 kasan_mempool_unpoison_pages(struct page *page, unsigned int order) argument 357 kasan_poison_pages(struct page *page, unsigned int order, bool init) argument 359 kasan_unpoison_pages(struct page *page, unsigned int order, bool init) argument 398 kasan_mempool_poison_pages(struct page *page, unsigned int order) argument 402 kasan_mempool_unpoison_pages(struct page *page, unsigned int order) argument [all...] |
/linux-master/arch/arm/mm/ |
H A D | flush.c | 128 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, argument 140 flush_pfn_alias(page_to_pfn(page), uaddr); 149 flush_icache_alias(page_to_pfn(page), uaddr, len); 159 void flush_ptrace_access(struct vm_area_struct *vma, struct page *page, argument 167 __flush_ptrace_access(page, uaddr, kaddr, len, flags); 170 void flush_uprobe_xol_access(struct page *page, unsigned long uaddr, argument 175 __flush_ptrace_access(page, uadd 185 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long uaddr, void *dst, const void *src, unsigned long len) argument 367 flush_dcache_page(struct page *page) argument 382 __flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument [all...] |
H A D | dma-mapping.c | 28 #include <asm/page.h> 55 struct page *page; member in struct:arm_dma_free_args 64 struct page **ret_page); 107 static void __dma_clear_buffer(struct page *page, size_t size, int coherent_flag) argument 113 if (PageHighMem(page)) { 114 phys_addr_t base = __pfn_to_phys(page_to_pfn(page)); 117 void *ptr = kmap_atomic(page); 122 page 145 struct page *page, *p, *e; local 166 __dma_free_buffer(struct page *page, size_t size) argument 204 struct page *page; local 312 struct page *page = virt_to_page((void *)addr); local 319 __dma_remap(struct page *page, size_t size, pgprot_t prot) argument 332 struct page *page; local 398 struct page *page; local 426 __free_from_contiguous(struct device *dev, struct page *page, void *cpu_addr, size_t size, bool want_vaddr) argument 449 struct page *page; local 538 struct page *page = NULL; local 609 struct page *page = phys_to_page(dma_to_phys(dev, handle)); local 627 dma_cache_maint_page(struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, void (*op)(const void *, size_t, int)) argument 679 __dma_page_cpu_to_dev(struct page *page, unsigned long off, size_t size, enum dma_data_direction dir) argument 695 __dma_page_dev_to_cpu(struct page *page, unsigned long off, size_t size, enum dma_data_direction dir) argument 869 struct page *page; local 1020 struct page *page; local 1044 struct page *page; local 1368 arm_iommu_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, unsigned long attrs) argument 1410 struct page *page; local 1488 struct page *page; local 1503 struct page *page; local [all...] |
/linux-master/fs/ocfs2/ |
H A D | mmap.c | 42 vma, vmf->page, vmf->pgoff); 47 struct buffer_head *di_bh, struct page *page) 53 loff_t pos = page_offset(page); 56 struct page *locked_page = NULL; 63 * There are cases that lead to the page no longer belonging to the 70 * then re-extended the file. We'll re-check the page mapping after 71 * taking the page lock inside of ocfs2_write_begin_nolock(). 75 if ((page->mapping != inode->i_mapping) || 76 (!PageUptodate(page)) || 46 __ocfs2_page_mkwrite(struct file *file, struct buffer_head *di_bh, struct page *page) argument 115 struct page *page = vmf->page; local [all...] |
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | main.c | 50 * from the input list, and made available for the page allocator. SECS pages 59 struct sgx_epc_page *page; local 68 page = list_first_entry(dirty_page_list, struct sgx_epc_page, list); 71 * Checking page->poison without holding the node->lock 74 * called for a page that sgx_free_epc_page() will put onto 77 if (page->poison) { 78 struct sgx_epc_section *section = &sgx_epc_sections[page->section]; 82 list_move(&page->list, &node->sgx_poison_page_list); 88 ret = __eremove(sgx_get_epc_virt_addr(page)); 91 * page i 111 struct sgx_encl_page *page = epc_page->owner; local 143 struct sgx_encl_page *page = epc_page->owner; local 444 struct sgx_epc_page *page = NULL; local 475 struct sgx_epc_page *page; local 506 sgx_mark_page_reclaimable(struct sgx_epc_page *page) argument 524 sgx_unmark_page_reclaimable(struct sgx_epc_page *page) argument 561 struct sgx_epc_page *page; local 602 sgx_free_epc_page(struct sgx_epc_page *page) argument 678 struct sgx_epc_page *page = sgx_paddr_to_page(pfn << PAGE_SHIFT); local [all...] |
/linux-master/drivers/hwmon/pmbus/ |
H A D | stpddc60.c | 85 static int stpddc60_read_byte_data(struct i2c_client *client, int page, int reg) argument 89 if (page > 0) 109 static int stpddc60_read_word_data(struct i2c_client *client, int page, argument 114 if (page > 0) 119 ret = pmbus_read_word_data(client, page, phase, 127 ret = pmbus_read_word_data(client, page, phase, reg); 145 static int stpddc60_write_word_data(struct i2c_client *client, int page, argument 151 if (page > 0) 156 ret = pmbus_read_word_data(client, page, 0xff, 161 ret = pmbus_write_byte_data(client, page, [all...] |
H A D | ltc3815.c | 27 static int ltc3815_read_byte_data(struct i2c_client *client, int page, int reg) argument 48 static int ltc3815_write_byte(struct i2c_client *client, int page, u8 reg) argument 72 static int ltc3815_read_word_data(struct i2c_client *client, int page, argument 79 ret = pmbus_read_word_data(client, page, phase, 83 ret = pmbus_read_word_data(client, page, phase, 87 ret = pmbus_read_word_data(client, page, phase, 91 ret = pmbus_read_word_data(client, page, phase, 95 ret = pmbus_read_word_data(client, page, phase, 112 static int ltc3815_write_word_data(struct i2c_client *client, int page, argument 119 ret = pmbus_write_word_data(client, page, [all...] |
/linux-master/fs/btrfs/ |
H A D | compression.h | 17 struct page; 48 struct page **compressed_pages; 89 u64 start, struct page **pages, 93 int btrfs_decompress(int type, const u8 *data_in, struct page *dest_page, 99 struct page **compressed_pages, 107 struct page *btrfs_alloc_compr_page(void); 108 void btrfs_free_compr_page(struct page *page); 153 u64 start, struct page **pages, unsigned long *out_pages, 157 struct page *dest_pag [all...] |
/linux-master/arch/x86/kernel/ |
H A D | espfix_64.c | 22 * This file sets up the ministacks and the related page tables. The 61 /* Page allocation bitmap - each page serves ESPFIX_STACKS_PER_PAGE CPUs */ 73 * we have to account for some amount of padding at the end of each page. 77 unsigned long page, slot; local 80 page = (cpu / ESPFIX_STACKS_PER_PAGE) ^ page_random; 82 addr = (page << PAGE_SHIFT) + (slot * ESPFIX_STACK_SIZE); 113 /* Install the espfix pud into the kernel page directory */ 127 unsigned int page; local 145 page = cpu/ESPFIX_STACKS_PER_PAGE; 148 stack_page = READ_ONCE(espfix_pages[page]); 165 struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0); local 177 struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0); local [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, 54 * have to bother with retries or mapping a dummy page to insure we 59 int qib_map_page(struct pci_dev *hwdev, struct page *page, dma_addr_t *daddr) argument 63 phys = dma_map_page(&hwdev->dev, page, 0, PAGE_SIZE, DMA_FROM_DEVICE); 69 phys = dma_map_page(&hwdev->dev, page, 0, PAGE_SIZE, 74 * FIXME: If we get 0 again, we should keep this page, 75 * map another, then free the 0 page. 84 * @start_page: the start page 86 * @p: the output page structure [all...] |
/linux-master/arch/arc/include/asm/ |
H A D | page.h | 8 #include <uapi/asm/page.h> 29 struct page; 33 void copy_user_highpage(struct page *to, struct page *from, 35 void clear_user_page(void *to, unsigned long u_vaddr, struct page *page); 85 typedef struct page *pgtable_t;
|
/linux-master/arch/csky/include/asm/ |
H A D | page.h | 11 * PAGE_SHIFT determines the page size: 4KB 43 #define clear_page(page) memset((page), 0, PAGE_SIZE) 46 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) 49 struct page; 51 #include <abi/page.h> 60 typedef struct page *pgtable_t;
|