/linux-master/fs/f2fs/ |
H A D | node.c | 107 * free memory is lower than watermark or cached page count 108 * exceed threshold, deny caching compress page. 123 static void clear_node_page_dirty(struct page *page) argument 125 if (PageDirty(page)) { 126 f2fs_clear_page_cache_dirty_tag(page); 127 clear_page_dirty_for_io(page); 128 dec_page_count(F2FS_P_SB(page), F2FS_DIRTY_NODES); 130 ClearPageUptodate(page); 133 static struct page *get_current_nat_pag 313 f2fs_in_warm_node_list(struct f2fs_sb_info *sbi, struct page *page) argument 327 f2fs_add_fsync_node_entry(struct f2fs_sb_info *sbi, struct page *page) argument 351 f2fs_del_fsync_node_entry(struct f2fs_sb_info *sbi, struct page *page) argument 554 struct page *page = NULL; local 937 struct page *page; local 976 struct page *page; local 1122 struct page *page; local 1311 struct page *page; local 1370 read_node_page(struct page *page, blk_opf_t op_flags) argument 1440 struct page *page; local 1516 struct page *page; local 1565 struct page *page = &fbatch.folios[i]->page; local 1606 __write_node_page(struct page *page, bool atomic, bool *submitted, struct writeback_control *wbc, bool do_balance, enum iostat_type io_type, unsigned int *seq_id) argument 1756 f2fs_write_node_page(struct page *page, struct writeback_control *wbc) argument 1791 struct page *page = &fbatch.folios[i]->page; local 1907 flush_dirty_inode(struct page *page) argument 1938 struct page *page = &fbatch.folios[i]->page; local 1992 struct page *page = &fbatch.folios[i]->page; local 2096 struct page *page; local 2515 struct page *page = get_current_nat_page(sbi, nid); local 2697 f2fs_recover_inline_xattr(struct inode *inode, struct page *page) argument 2734 f2fs_recover_xattr_data(struct inode *inode, struct page *page) argument 2782 f2fs_recover_inode_page(struct f2fs_sb_info *sbi, struct page *page) argument 2873 struct page *page = f2fs_get_tmp_page(sbi, idx); local 2968 update_nat_bits(struct f2fs_sb_info *sbi, nid_t start_nid, struct page *page) argument 3029 struct page *page = NULL; local 3182 struct page *page; local [all...] |
H A D | data.c | 51 bool f2fs_is_cp_guaranteed(struct page *page) argument 53 struct address_space *mapping = page->mapping; 69 page_private_gcing(page)) 74 static enum count_type __read_io_type(struct page *page) argument 76 struct address_space *mapping = page_file_mapping(page); 133 * things for each compressed page here: call f2fs_end_read_compressed_page() 136 * release the bio's reference to the decompress_io_ctx of the page's cluster. 145 struct page *pag local 194 struct page *page = bv->bv_page; local 248 struct page *page = bv->bv_page; local 338 struct page *page = bvec->bv_page; local 546 __has_merged_page(struct bio *bio, struct inode *inode, struct page *page, nid_t ino) argument 642 __submit_merged_write_cond(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page, nid_t ino, enum page_type type, bool force) argument 672 f2fs_submit_merged_write_cond(struct f2fs_sb_info *sbi, struct inode *inode, struct page *page, nid_t ino, enum page_type type) argument 693 struct page *page = fio->encrypted_page ? local 757 add_bio_entry(struct f2fs_sb_info *sbi, struct bio *bio, struct page *page, enum temp_type temp) argument 781 add_ipu_page(struct f2fs_io_info *fio, struct bio **bio, struct page *page) argument 829 f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi, struct bio **bio, struct page *page) argument 890 struct page *page = fio->encrypted_page ? local 1088 f2fs_submit_page_read(struct inode *inode, struct page *page, block_t blkaddr, blk_opf_t op_flags, bool for_write) argument 1211 struct page *page; local 1286 struct page *page; local 1317 struct page *page; local 1345 struct page *page; local 1826 struct page *page; local 2203 struct page *page = cc->rpages[i]; local 2276 struct page *page = dic->cpages[i]; local 2502 struct page *mpage, *page; local 2641 struct page *page = fio->page; local 2756 f2fs_write_single_data_page(struct page *page, int *submitted, struct bio **bio, sector_t *last_block, struct writeback_control *wbc, enum iostat_type io_type, int compr_blocks, bool allow_balance) argument 2929 f2fs_write_data_page(struct page *page, struct writeback_control *wbc) argument 3063 struct page *page = pages[i]; local 3364 prepare_write_begin(struct f2fs_sb_info *sbi, struct page *page, loff_t pos, unsigned len, block_t *blk_addr, bool *node_changed) argument 3504 prepare_atomic_write_begin(struct f2fs_sb_info *sbi, struct page *page, loff_t pos, unsigned int len, block_t *blk_addr, bool *node_changed, bool *use_cow) argument 3552 struct page *page = NULL; local 3678 f2fs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument 3882 struct page *page; local 4106 f2fs_clear_page_cache_dirty_tag(struct page *page) argument [all...] |
/linux-master/drivers/nvme/target/ |
H A D | configfs.c | 64 static ssize_t nvmet_addr_adrfam_show(struct config_item *item, char *page) argument 71 return snprintf(page, PAGE_SIZE, "%s\n", 75 return snprintf(page, PAGE_SIZE, "\n"); 79 const char *page, size_t count) 88 if (sysfs_streq(page, nvmet_addr_family[i].name)) 92 pr_err("Invalid value '%s' for adrfam\n", page); 103 char *page) 107 return snprintf(page, PAGE_SIZE, "%d\n", le16_to_cpu(portid)); 111 const char *page, size_t count) 116 if (kstrtou16(page, 78 nvmet_addr_adrfam_store(struct config_item *item, const char *page, size_t count) argument 102 nvmet_addr_portid_show(struct config_item *item, char *page) argument 110 nvmet_addr_portid_store(struct config_item *item, const char *page, size_t count) argument 130 nvmet_addr_traddr_show(struct config_item *item, char *page) argument 138 nvmet_addr_traddr_store(struct config_item *item, const char *page, size_t count) argument 169 nvmet_addr_treq_show(struct config_item *item, char *page) argument 183 nvmet_addr_treq_store(struct config_item *item, const char *page, size_t count) argument 223 nvmet_addr_trsvcid_show(struct config_item *item, char *page) argument 231 nvmet_addr_trsvcid_store(struct config_item *item, const char *page, size_t count) argument 250 nvmet_param_inline_data_size_show(struct config_item *item, char *page) argument 258 nvmet_param_inline_data_size_store(struct config_item *item, const char *page, size_t count) argument 276 nvmet_param_max_queue_size_show(struct config_item *item, char *page) argument 284 nvmet_param_max_queue_size_store(struct config_item *item, const char *page, size_t count) argument 303 nvmet_param_pi_enable_show(struct config_item *item, char *page) argument 311 nvmet_param_pi_enable_store(struct config_item *item, const char *page, size_t count) argument 330 nvmet_addr_trtype_show(struct config_item *item, char *page) argument 357 nvmet_addr_trtype_store(struct config_item *item, const char *page, size_t count) argument 396 nvmet_addr_tsas_show(struct config_item *item, char *page) argument 416 nvmet_addr_tsas_store(struct config_item *item, const char *page, size_t count) argument 475 nvmet_ns_device_path_show(struct config_item *item, char *page) argument 480 nvmet_ns_device_path_store(struct config_item *item, const char *page, size_t count) argument 515 nvmet_ns_p2pmem_show(struct config_item *item, char *page) argument 522 nvmet_ns_p2pmem_store(struct config_item *item, const char *page, size_t count) argument 556 nvmet_ns_device_uuid_show(struct config_item *item, char *page) argument 561 nvmet_ns_device_uuid_store(struct config_item *item, const char *page, size_t count) argument 584 nvmet_ns_device_nguid_show(struct config_item *item, char *page) argument 589 nvmet_ns_device_nguid_store(struct config_item *item, const char *page, size_t count) argument 630 nvmet_ns_ana_grpid_show(struct config_item *item, char *page) argument 635 nvmet_ns_ana_grpid_store(struct config_item *item, const char *page, size_t count) argument 664 nvmet_ns_enable_show(struct config_item *item, char *page) argument 669 nvmet_ns_enable_store(struct config_item *item, const char *page, size_t count) argument 697 nvmet_ns_buffered_io_show(struct config_item *item, char *page) argument 702 nvmet_ns_buffered_io_store(struct config_item *item, const char *page, size_t count) argument 725 nvmet_ns_revalidate_size_store(struct config_item *item, const char *page, size_t count) argument 836 nvmet_passthru_device_path_show(struct config_item *item, char *page) argument 844 nvmet_passthru_device_path_store(struct config_item *item, const char *page, size_t count) argument 877 nvmet_passthru_enable_show(struct config_item *item, char *page) argument 885 nvmet_passthru_enable_store(struct config_item *item, const char *page, size_t count) argument 904 nvmet_passthru_admin_timeout_show(struct config_item *item, char *page) argument 910 nvmet_passthru_admin_timeout_store(struct config_item *item, const char *page, size_t count) argument 923 nvmet_passthru_io_timeout_show(struct config_item *item, char *page) argument 929 nvmet_passthru_io_timeout_store(struct config_item *item, const char *page, size_t count) argument 942 nvmet_passthru_clear_ids_show(struct config_item *item, char *page) argument 948 nvmet_passthru_clear_ids_store(struct config_item *item, const char *page, size_t count) argument 1145 nvmet_subsys_attr_allow_any_host_show(struct config_item *item, char *page) argument 1152 nvmet_subsys_attr_allow_any_host_store(struct config_item *item, const char *page, size_t count) argument 1181 nvmet_subsys_attr_version_show(struct config_item *item, char *page) argument 1198 nvmet_subsys_attr_version_store_locked(struct nvmet_subsys *subsys, const char *page, size_t count) argument 1230 nvmet_subsys_attr_version_store(struct config_item *item, const char *page, size_t count) argument 1252 nvmet_subsys_attr_serial_show(struct config_item *item, char *page) argument 1262 nvmet_subsys_attr_serial_store_locked(struct nvmet_subsys *subsys, const char *page, size_t count) argument 1291 nvmet_subsys_attr_serial_store(struct config_item *item, const char *page, size_t count) argument 1307 nvmet_subsys_attr_cntlid_min_show(struct config_item *item, char *page) argument 1313 nvmet_subsys_attr_cntlid_min_store(struct config_item *item, const char *page, size_t cnt) argument 1337 nvmet_subsys_attr_cntlid_max_show(struct config_item *item, char *page) argument 1343 nvmet_subsys_attr_cntlid_max_store(struct config_item *item, const char *page, size_t cnt) argument 1367 nvmet_subsys_attr_model_show(struct config_item *item, char *page) argument 1375 nvmet_subsys_attr_model_store_locked(struct nvmet_subsys *subsys, const char *page, size_t count) argument 1410 nvmet_subsys_attr_model_store(struct config_item *item, const char *page, size_t count) argument 1426 nvmet_subsys_attr_ieee_oui_show(struct config_item *item, char *page) argument 1434 nvmet_subsys_attr_ieee_oui_store_locked(struct nvmet_subsys *subsys, const char *page, size_t count) argument 1458 nvmet_subsys_attr_ieee_oui_store(struct config_item *item, const char *page, size_t count) argument 1474 nvmet_subsys_attr_firmware_show(struct config_item *item, char *page) argument 1482 nvmet_subsys_attr_firmware_store_locked(struct nvmet_subsys *subsys, const char *page, size_t count) argument 1520 nvmet_subsys_attr_firmware_store(struct config_item *item, const char *page, size_t count) argument 1537 nvmet_subsys_attr_pi_enable_show(struct config_item *item, char *page) argument 1543 nvmet_subsys_attr_pi_enable_store(struct config_item *item, const char *page, size_t count) argument 1558 nvmet_subsys_attr_qid_max_show(struct config_item *item, char *page) argument 1564 nvmet_subsys_attr_qid_max_store(struct config_item *item, const char *page, size_t cnt) argument 1670 nvmet_referral_enable_show(struct config_item *item, char *page) argument 1676 nvmet_referral_enable_store(struct config_item *item, const char *page, size_t count) argument 1772 nvmet_ana_group_ana_state_show(struct config_item *item, char *page) argument 1787 nvmet_ana_group_ana_state_store(struct config_item *item, const char *page, size_t count) argument 2015 nvmet_host_dhchap_key_show(struct config_item *item, char *page) argument 2031 nvmet_host_dhchap_key_store(struct config_item *item, const char *page, size_t count) argument 2048 nvmet_host_dhchap_ctrl_key_show(struct config_item *item, char *page) argument 2064 nvmet_host_dhchap_ctrl_key_store(struct config_item *item, const char *page, size_t count) argument 2081 nvmet_host_dhchap_hash_show(struct config_item *item, char *page) argument 2090 nvmet_host_dhchap_hash_store(struct config_item *item, const char *page, size_t count) argument 2107 nvmet_host_dhchap_dhgroup_show(struct config_item *item, char *page) argument 2116 nvmet_host_dhchap_dhgroup_store(struct config_item *item, const char *page, size_t count) argument 2199 nvmet_root_discovery_nqn_show(struct config_item *item, char *page) argument 2205 nvmet_root_discovery_nqn_store(struct config_item *item, const char *page, size_t count) argument [all...] |
/linux-master/fs/crypto/ |
H A D | crypto.c | 50 struct page *fscrypt_alloc_bounce_page(gfp_t gfp_flags) 55 * page pool, but it didn't set needs_bounce_pages. 63 * fscrypt_free_bounce_page() - free a ciphertext bounce page 64 * @bounce_page: the bounce page to free, or NULL 66 * Free a bounce page that was allocated by fscrypt_encrypt_pagecache_blocks(), 69 void fscrypt_free_bounce_page(struct page *bounce_page) 110 struct page *src_page, struct page *dest_page, 156 * fscrypt_encrypt_pagecache_blocks() - Encrypt data from a pagecache page 157 * @page 180 fscrypt_encrypt_pagecache_blocks(struct page *page, unsigned int len, unsigned int offs, gfp_t gfp_flags) argument 240 fscrypt_encrypt_block_inplace(const struct inode *inode, struct page *page, unsigned int len, unsigned int offs, u64 lblk_num, gfp_t gfp_flags) argument 285 struct page *page = folio_page(folio, i >> PAGE_SHIFT); local 315 fscrypt_decrypt_block_inplace(const struct inode *inode, struct page *page, unsigned int len, unsigned int offs, u64 lblk_num) argument [all...] |
/linux-master/arch/mips/mm/ |
H A D | init.c | 52 * any price. Since page is never written to after the initialization we 65 struct page *page; local 76 page = virt_to_page((void *)empty_zero_page); 77 split_page(page, order); 78 for (i = 0; i < (1 << order); i++, page++) 79 mark_page_reserved(page); 84 static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot) argument 93 BUG_ON(folio_test_dcache_dirty(page_folio(page))); 139 kmap_coherent(struct page *page, unsigned long addr) argument 144 kmap_noncoherent(struct page *page, unsigned long addr) argument 196 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 216 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 454 struct page *page = pfn_to_page(tmp); local 492 struct page *page = pfn_to_page(pfn); local [all...] |
/linux-master/drivers/hwmon/pmbus/ |
H A D | pmbus.c | 26 * Find sensor groups and status registers on each page. 31 int page; local 33 /* Sensors detected on page 0 only */ 70 for (page = 0; page < info->pages; page++) { 71 if (pmbus_check_word_register(client, page, PMBUS_READ_VOUT)) { 72 info->func[page] |= PMBUS_HAVE_VOUT; 73 if (pmbus_check_byte_register(client, page, 75 info->func[page] | 104 int page; local [all...] |
H A D | max15301.c | 69 static int max15301_read_word_data(struct i2c_client *client, int page, argument 76 if (page > 0) 83 ret = pmbus_read_word_data(client, page, phase, reg); 89 static int max15301_read_byte_data(struct i2c_client *client, int page, int reg) argument 95 if (page > 0) 99 ret = pmbus_read_byte_data(client, page, reg); 105 static int max15301_write_word_data(struct i2c_client *client, int page, int reg, argument 112 if (page > 0) 119 ret = pmbus_write_word_data(client, page, reg, word); 125 static int max15301_write_byte(struct i2c_client *client, int page, u argument [all...] |
H A D | max16064.c | 18 static int max16064_read_word_data(struct i2c_client *client, int page, argument 25 ret = pmbus_read_word_data(client, page, phase, 29 ret = pmbus_read_word_data(client, page, phase, 43 static int max16064_write_word_data(struct i2c_client *client, int page, argument 50 ret = pmbus_write_word_data(client, page, 54 ret = pmbus_write_word_data(client, page,
|
/linux-master/arch/mips/include/asm/ |
H A D | tlbflush.h | 12 * - flush_tlb_page(vma, vmaddr) flushes one page 22 unsigned long page); 44 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/linux-master/tools/testing/scatterlist/linux/ |
H A D | mm.h | 45 static inline unsigned long page_to_phys(struct page *page) argument 52 #define page_to_pfn(page) ((unsigned long)(page) / PAGE_SIZE) 54 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) 79 static inline void *kmap(struct page *page) argument 86 static inline void *kmap_atomic(struct page *page) argument 108 free_page(unsigned long page) argument [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | tlbflush.h | 15 * - flush_tlb_page(vma, vmaddr) flushes one page 25 extern void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 43 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/linux-master/mm/ |
H A D | swap_cgroup.c | 10 struct page **map; 41 struct page *page; local 48 page = alloc_page(GFP_KERNEL | __GFP_ZERO); 49 if (!page) 51 ctrl->map[idx] = page; 68 struct page *mappage; 206 struct page **map; 223 struct page *page local [all...] |
H A D | internal.h | 128 * @addr: The user virtual address the first page is mapped at. 130 * @pte: Page table entry for the first page. 147 * start_ptep must map any page of the folio. max_nr must be at least one and 148 * must be limited by the caller so scanning cannot exceed a single page table. 248 * cannot exceed a single page table. 360 * Turn a non-refcounted page (->_refcount == 0) into refcounted with 363 static inline void set_page_refcounted(struct page *page) 365 VM_BUG_ON_PAGE(PageTail(page), page); [all...] |
/linux-master/drivers/net/thunderbolt/ |
H A D | trace.h | 26 TP_PROTO(unsigned int index, const void *page, dma_addr_t phys, 28 TP_ARGS(index, page, phys, dir), 31 __field(const void *, page) 37 __entry->page = page; 41 TP_printk("index=%u page=%p phys=%pad dir=%s", 42 __entry->index, __entry->page, &__entry->phys, 47 TP_PROTO(unsigned int index, const void *page, dma_addr_t phys, 49 TP_ARGS(index, page, phys, dir) 53 TP_PROTO(unsigned int index, const void *page, dma_addr_ [all...] |
/linux-master/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 17 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page 20 * - flush_dcache_page(pg) flushes(wback&invalidates) a page for dcache 71 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, 75 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/arch/arm64/mm/ |
H A D | pageattr.c | 27 * mapped at page granularity, so that it is possible to 30 * KFENCE pool requires page-granular mapping if initialized late. 84 * mappings into page mappings may cause TLB conflicts. This means 89 * Those are guaranteed to consist entirely of page mappings, and 165 int set_direct_map_invalid_noflush(struct page *page) argument 176 (unsigned long)page_address(page), 180 int set_direct_map_default_noflush(struct page *page) argument 191 (unsigned long)page_address(page), 196 __kernel_map_pages(struct page *page, int numpages, int enable) argument 213 kernel_page_present(struct page *page) argument [all...] |
/linux-master/arch/arc/mm/ |
H A D | dma.c | 18 void arch_dma_prep_coherent(struct page *page, size_t size) argument 21 * Evict any existing L1 and/or L2 lines for the backing page 22 * in case it was used earlier as a normal "cached" page. 30 dma_cache_wback_inv(page_to_phys(page), size);
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_memfree.c | 40 #include <asm/page.h> 61 } page[]; member in struct:mthca_user_db_table 109 struct page *page; local 115 page = alloc_pages(gfp_mask | __GFP_ZERO, order); 116 if (!page) 119 sg_set_page(mem, page, PAGE_SIZE << order, 0); 283 struct page *page = NULL; local 306 * so if we found the page, dma_handl 439 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page) argument 575 struct mthca_db_page *page; local 670 struct mthca_db_page *page; local [all...] |
/linux-master/drivers/net/dsa/b53/ |
H A D | b53_mdio.c | 42 static int b53_mdio_op(struct b53_device *dev, u8 page, u8 reg, u16 op) argument 49 if (dev->current_page != page) { 50 /* set page number */ 51 v = (page << 8) | REG_MII_PAGE_ENABLE; 56 dev->current_page = page; 80 static int b53_mdio_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) argument 85 ret = b53_mdio_op(dev, page, reg, REG_MII_ADDR_READ); 95 static int b53_mdio_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) argument 100 ret = b53_mdio_op(dev, page, reg, REG_MII_ADDR_READ); 109 static int b53_mdio_read32(struct b53_device *dev, u8 page, u argument 125 b53_mdio_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 147 b53_mdio_read64(struct b53_device *dev, u8 page, u8 reg, u64 *val) argument 169 b53_mdio_write8(struct b53_device *dev, u8 page, u8 reg, u8 value) argument 182 b53_mdio_write16(struct b53_device *dev, u8 page, u8 reg, u16 value) argument 196 b53_mdio_write32(struct b53_device *dev, u8 page, u8 reg, u32 value) argument 215 b53_mdio_write48(struct b53_device *dev, u8 page, u8 reg, u64 value) argument 234 b53_mdio_write64(struct b53_device *dev, u8 page, u8 reg, u64 value) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | pageattr.c | 14 #include <asm/page.h> 26 * Updates the attributes of a page atomically. 29 * attributes of a page currently being executed or accessed. 89 * On hash, the linear mapping is not in the Linux page table so 108 void __kernel_map_pages(struct page *page, int numpages, int enable) argument 111 unsigned long addr = (unsigned long)page_address(page); 113 if (PageHighMem(page)) 117 err = hash__kernel_map_pages(page, numpages, enable);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/dma/ |
H A D | usergf119.c | 76 u32 kind, page; local 94 "create gf100 dma vers %d page %d kind %02x\n", 95 args->v0.version, args->v0.page, args->v0.kind); 97 page = args->v0.page; 102 page = GF119_DMA_V0_PAGE_SP; 105 page = GF119_DMA_V0_PAGE_LP; 110 if (page > 1) 112 dmaobj->flags0 = (kind << 20) | (page << 6);
|
/linux-master/include/xen/ |
H A D | xen-front-pgdir-shbuf.h | 4 * Xen frontend/backend page directory based shared buffer 26 * of grant references for the page directory and the pages 30 * of grant references for the page directory itself as grant 39 * Number of pages for the shared buffer itself (excluding the page 47 struct page **pages; 64 struct page **pages;
|
/linux-master/arch/powerpc/include/asm/ |
H A D | async_tx.h | 13 struct page **dst_lst, int dst_cnt, struct page **src_lst,
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.h | 42 unsigned long nouveau_dmem_page_addr(struct page *page);
|
/linux-master/include/linux/mlx5/ |
H A D | rsc_dump.h | 50 struct page *page, int *size);
|