/linux-master/arch/s390/include/asm/ |
H A D | page-states.h | 38 static __always_inline void __set_page_state(void *addr, unsigned long num_pages, unsigned char cmd) argument 42 while (num_pages--) { 48 static inline void __set_page_unused(void *addr, unsigned long num_pages) argument 50 __set_page_state(addr, num_pages, ESSA_SET_UNUSED); 53 static inline void __set_page_stable_dat(void *addr, unsigned long num_pages) argument 55 __set_page_state(addr, num_pages, ESSA_SET_STABLE); 58 static inline void __set_page_stable_nodat(void *addr, unsigned long num_pages) argument 60 __set_page_state(addr, num_pages, ESSA_SET_STABLE_NODAT); 63 static inline void __arch_set_page_nodat(void *addr, unsigned long num_pages) argument 68 __set_page_stable_dat(addr, num_pages); 73 __arch_set_page_dat(void *addr, unsigned long num_pages) argument [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | mremap_dontunmap.c | 45 unsigned long num_pages = 1; local 46 void *source_mapping = mmap(NULL, num_pages * page_size, PROT_NONE, 53 mremap(source_mapping, num_pages * page_size, num_pages * page_size, 58 BUG_ON(munmap(dest_mapping, num_pages * page_size) == -1, 62 BUG_ON(munmap(source_mapping, num_pages * page_size) == -1, 78 unsigned long num_pages = size / page_size; local 82 for (i = 0; i < num_pages; ++i) { 97 unsigned long num_pages = 5; local 100 mmap(NULL, num_pages * page_siz 131 unsigned long num_pages = 5; local 185 unsigned long num_pages = 5; local 245 unsigned long num_pages = 10; local [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, argument 43 unpin_user_pages_dirty_lock(p, num_pages, dirty); 85 * @num_pages: the number of pages 90 * now, num_pages is always 1, but that will probably change at some point 94 int qib_get_user_pages(unsigned long start_page, size_t num_pages, argument 102 locked = atomic64_add_return(num_pages, ¤t->mm->pinned_vm); 110 for (got = 0; got < num_pages; got += ret) { 112 num_pages - got, 126 atomic64_sub(num_pages, ¤t->mm->pinned_vm); 130 void qib_release_user_pages(struct page **p, size_t num_pages) argument [all...] |
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_reg.h | 39 u32 num_pages; member in struct:svga_guest_mem_descriptor
|
H A D | vmwgfx_gmr.c | 39 unsigned long num_pages, 47 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); 48 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; 58 define_cmd.numPages = num_pages; 73 while (num_pages > 0) { 74 unsigned long nr = min_t(unsigned long, num_pages, VMW_PPN_PER_REMAP); 94 num_pages -= nr; 128 unsigned long num_pages, 141 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_i 37 vmw_gmr2_bind(struct vmw_private *dev_priv, struct vmw_piter *iter, unsigned long num_pages, int gmr_id) argument 126 vmw_gmr_bind(struct vmw_private *dev_priv, const struct vmw_sg_table *vsgt, unsigned long num_pages, int gmr_id) argument [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | set_memory.h | 21 int num_pages)) 25 int num_pages = PAGE_ALIGN(end - start) >> PAGE_SHIFT; local 27 return set_memory(start, num_pages); 37 int num_pages)) 19 set_kernel_memory(char *startp, char *endp, int (*set_memory)(unsigned long start, int num_pages)) argument 35 set_kernel_memory(char *startp, char *endp, int (*set_memory)(unsigned long start, int num_pages)) argument
|
/linux-master/include/xen/ |
H A D | xen-front-pgdir-shbuf.h | 42 int num_pages; member in struct:xen_front_pgdir_shbuf 62 int num_pages; member in struct:xen_front_pgdir_shbuf_cfg
|
/linux-master/drivers/gpu/drm/gma500/ |
H A D | mmu.h | 69 uint32_t num_pages); 73 uint32_t num_pages, int type); 78 unsigned long address, uint32_t num_pages, 82 unsigned long address, uint32_t num_pages,
|
/linux-master/tools/testing/scatterlist/ |
H A D | main.c | 10 unsigned num_pages; member in struct:test 40 printf("%u input PFNs:", test->num_pages); 41 for (i = 0; i < test->num_pages; i++) 87 int left_pages = test->pfn_app ? test->num_pages : 0; 92 set_pages(pages, test->pfn, test->num_pages); 96 &append, pages, test->num_pages, 0, test->size, 100 &append.sgt, pages, test->num_pages, 0, 109 set_pages(pages, test->pfn_app, test->num_pages); 111 &append, pages, test->num_pages, 0, test->size,
|
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_pool.c | 168 unsigned int num_pages = last - first; local 170 if (!num_pages) 177 return set_pages_array_wc(first, num_pages); 179 return set_pages_array_uc(first, num_pages); 214 unsigned int num_pages) 220 dma_unmap_page(pool->dev, dma_addr, (long)num_pages << PAGE_SHIFT, 227 unsigned int i, num_pages = 1 << pt->order; local 229 for (i = 0; i < num_pages; ++i) { 324 unsigned int num_pages; local 336 num_pages 213 ttm_pool_unmap(struct ttm_pool *pool, dma_addr_t dma_addr, unsigned int num_pages) argument 358 ttm_pool_page_allocated(struct ttm_pool *pool, unsigned int order, struct page *p, dma_addr_t **dma_addr, unsigned long *num_pages, struct page ***pages) argument 431 pgoff_t num_pages = tt->num_pages; local 650 unsigned long num_pages = atomic_long_read(&allocated_pages); local 787 ttm_pool_mgr_init(unsigned long num_pages) argument [all...] |
H A D | ttm_tt.c | 115 ttm->pages = kvcalloc(ttm->num_pages, sizeof(void*), GFP_KERNEL); 124 ttm->pages = kvcalloc(ttm->num_pages, sizeof(*ttm->pages) + 129 ttm->dma_address = (void *)(ttm->pages + ttm->num_pages); 135 ttm->dma_address = kvcalloc(ttm->num_pages, sizeof(*ttm->dma_address), 155 ttm->num_pages = (PAGE_ALIGN(bo->base.size) >> PAGE_SHIFT) + extra_pages; 228 for (i = 0; i < ttm->num_pages; ++i) { 268 loff_t size = (loff_t)ttm->num_pages << PAGE_SHIFT; 284 for (i = 0; i < ttm->num_pages; ++i) { 304 return ttm->num_pages; 324 atomic_long_add(ttm->num_pages, 410 ttm_tt_mgr_init(unsigned long num_pages, unsigned long num_dma32_pages) argument [all...] |
H A D | ttm_device.c | 70 unsigned long num_pages, num_dma32; local 88 num_pages = ((u64)si.totalram * si.mem_unit) >> PAGE_SHIFT; 89 num_pages /= 2; 96 ttm_pool_mgr_init(num_pages); 97 ttm_tt_mgr_init(num_pages, num_dma32); 165 uint32_t num_pages; local 170 num_pages = PFN_UP(bo->base.size); 174 return num_pages;
|
/linux-master/include/drm/ |
H A D | drm_cache.h | 40 void drm_clflush_pages(struct page *pages[], unsigned long num_pages);
|
/linux-master/drivers/xen/ |
H A D | xen-front-pgdir-shbuf.c | 160 return DIV_ROUND_UP(buf->num_pages, XEN_NUM_GREFS_PER_PAGE); 187 buf->num_grefs = get_num_pages_dir(buf) + buf->num_pages; 208 unmap_ops = kcalloc(buf->num_pages, sizeof(*unmap_ops), 213 for (i = 0; i < buf->num_pages; i++) { 222 buf->num_pages); 224 for (i = 0; i < buf->num_pages; i++) { 253 map_ops = kcalloc(buf->num_pages, sizeof(*map_ops), GFP_KERNEL); 257 buf->backend_map_handles = kcalloc(buf->num_pages, 271 grefs_left = buf->num_pages; 295 ret = gnttab_map_refs(map_ops, NULL, buf->pages, buf->num_pages); [all...] |
/linux-master/drivers/gpu/drm/xen/ |
H A D | xen_drm_front_gem.c | 29 size_t num_pages; member in struct:xen_gem_object 48 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); 49 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, 93 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); 159 ret = xen_alloc_unpopulated_pages(xen_obj->num_pages, 163 xen_obj->num_pages, ret); 175 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); 212 xen_free_unpopulated_pages(xen_obj->num_pages, 240 xen_obj->pages, xen_obj->num_pages); 265 xen_obj->num_pages); [all...] |
/linux-master/net/ceph/ |
H A D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) argument 17 for (i = 0; i < num_pages; i++) { 26 void ceph_release_page_vector(struct page **pages, int num_pages) argument 30 for (i = 0; i < num_pages; i++) 39 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) argument 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); 47 for (i = 0; i < num_pages; i++) {
|
/linux-master/arch/x86/hyperv/ |
H A D | hv_proc.c | 23 int hv_call_deposit_pages(int node, u64 partition_id, u32 num_pages) argument 36 if (num_pages > HV_DEPOSIT_MAX) 38 if (!num_pages) 56 while (num_pages) { 58 order = 31 - __builtin_clz(num_pages); 74 num_pages -= counts[i];
|
/linux-master/drivers/firmware/efi/ |
H A D | memattr.c | 59 u64 in_size = in->num_pages << EFI_PAGE_SHIFT; 72 !PAGE_ALIGNED(in->num_pages << EFI_PAGE_SHIFT))) { 86 u64 md_size = md->num_pages << EFI_PAGE_SHIFT; 169 size = md.num_pages << EFI_PAGE_SHIFT;
|
/linux-master/arch/x86/platform/efi/ |
H A D | memmap.c | 118 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; 183 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; 192 md->num_pages = (m_end - md->phys_addr + 1) >> 199 md->num_pages = (end - md->phys_addr + 1) >> 205 md->num_pages = (m_start - md->phys_addr) >> 213 md->num_pages = (m_end - m_start + 1) >> 220 md->num_pages = (end - m_end) >> 227 md->num_pages = (m_start - md->phys_addr) >> 234 md->num_pages = (end - md->phys_addr + 1) >>
|
H A D | efi_32.c | 43 size = md->num_pages << PAGE_SHIFT; 51 set_memory_uc((unsigned long)va, md->num_pages); 82 int __init efi_setup_page_tables(unsigned long pa_memmap, unsigned num_pages) argument 139 set_memory_x(md->virt_addr, md->num_pages);
|
/linux-master/include/linux/ |
H A D | tee_drv.h | 56 * @num_pages: number of locked pages 72 size_t num_pages; member in struct:tee_shm 157 * @num_pages: Number of pages will be stored there 161 size_t *num_pages) 163 *num_pages = shm->num_pages; 160 tee_shm_get_pages(struct tee_shm *shm, size_t *num_pages) argument
|
/linux-master/drivers/tee/ |
H A D | tee_shm.c | 38 unpin_user_pages(shm->pages, shm->num_pages); 40 shm_put_kernel_pages(shm->pages, shm->num_pages); 210 size_t num_pages, 240 shm->num_pages = nr_pages; 277 size_t num_pages, off; local 305 num_pages = iov_iter_npages(iter, INT_MAX); 306 if (!num_pages) { 311 shm->pages = kcalloc(num_pages, sizeof(*shm->pages), GFP_KERNEL); 317 len = iov_iter_extract_pages(iter, &shm->pages, LONG_MAX, num_pages, 0, 329 shm_get_kernel_pages(shm->pages, num_pages); 206 tee_dyn_shm_alloc_helper(struct tee_shm *shm, size_t size, size_t align, int (*shm_register)(struct tee_context *ctx, struct tee_shm *shm, struct page **pages, size_t num_pages, unsigned long start)) argument [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_mem.c | 68 int i, num_pages = umem->num_pages; local 73 for (i = 0; num_pages > 0; i++) { 75 num_pages -= PAGES_PER_CHUNK; 365 int num_pages, num_chunks, i, rv = 0; local 371 num_pages = PAGE_ALIGN(start + len - first_page_va) >> PAGE_SHIFT; 372 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; 400 for (i = 0; num_pages > 0; i++) { 401 int nents = min_t(int, num_pages, PAGES_PER_CHUNK); 412 umem->num_pages [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_userptr.c | 92 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; local 94 unpin_user_pages(pvec, num_pages); 104 unsigned int num_pages; /* limited by sg_alloc_table_from_pages_segment */ local 107 if (overflows_type(obj->base.size >> PAGE_SHIFT, num_pages)) 110 num_pages = obj->base.size >> PAGE_SHIFT; 124 ret = sg_alloc_table_from_pages_segment(st, pvec, num_pages, 0, 125 num_pages << PAGE_SHIFT, 235 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; local 260 pvec = kvmalloc_array(num_pages, sizeof(struct page *), GFP_KERNEL); 268 while (pinned < num_pages) { [all...] |
/linux-master/fs/crypto/ |
H A D | bio.c | 56 int num_pages = 0; local 66 if (num_pages == 0) { 76 num_pages++; 80 if (num_pages == BIO_MAX_VECS || !len || 86 num_pages = 0;
|