/linux-master/arch/x86/xen/ |
H A D | setup.c | 513 unsigned long max_pages, limit; local 518 max_pages = limit; 532 max_pages = ret; 535 return min(max_pages, limit); 692 unsigned long max_pages; local 750 max_pages = xen_get_max_pages(); 753 max_pages += xen_foreach_remap_area(max_pfn, xen_count_remap_pages); 755 if (max_pages > max_pfn) 756 extra_pages += max_pages - max_pfn; 762 * Make sure we have no memory above max_pages, a [all...] |
/linux-master/crypto/ |
H A D | algif_hash.c | 70 size_t len, max_pages, npages; local 74 max_pages = min_t(size_t, ALG_MAX_PAGES, 111 npages = iov_iter_npages(&msg->msg_iter, max_pages);
|
/linux-master/drivers/base/ |
H A D | memory.c | 1103 * @max_pages: The maximum number of pages we'll have in this static memory 1112 * memory groups can be registered, or if max_pages is invalid (0). Otherwise, 1115 int memory_group_register_static(int nid, unsigned long max_pages) argument 1120 .max_pages = max_pages, 1124 if (!max_pages)
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 1989 uint64_t max_pages; local 2006 max_pages = READ_ONCE(max_svm_range_pages); 2007 _pages = min_not_zero(max_pages, min_pages); 2008 } while (cmpxchg(&max_svm_range_pages, max_pages, _pages) != max_pages); 2013 uint64_t max_pages, struct list_head *insert_list, 2020 max_pages, start, last); 2023 l = min(last, ALIGN_DOWN(start + max_pages, max_pages) - 1); 2012 svm_range_split_new(struct svm_range_list *svms, uint64_t start, uint64_t last, uint64_t max_pages, struct list_head *insert_list, struct list_head *update_list) argument
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 719 unsigned long max_pages; local 740 max_pages = vm->total >> PAGE_SHIFT; 742 for_each_prime_number_from(page_num, 1, max_pages) {
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_gem_gtt.c | 416 const unsigned long max_pages = local 418 const unsigned long max_step = max(int_sqrt(max_pages), 2UL); 431 for (npages = 1; npages <= max_pages; npages *= prime) { 638 const unsigned long max_pages = local 652 for_each_prime_number_from(size, 1, max_pages) {
|
H A D | i915_vma.c | 577 const unsigned int max_pages = 64; local 584 obj = i915_gem_object_create_internal(vm->i915, max_pages * PAGE_SIZE); 601 GEM_BUG_ON(max_offset > max_pages); 602 max_offset = max_pages - max_offset;
|
H A D | intel_memory_region.c | 60 unsigned long max_pages; local 66 max_pages = div64_u64(total, page_size); 69 for_each_prime_number_from(page_num, 1, max_pages) {
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_free_list.c | 91 free_list->max_pages = args->max_num_pages; 179 fw_data->max_pages = free_list->max_pages; 188 ((fw_data->max_pages - fw_data->current_pages) * 280 start_page = free_list->max_pages - free_list->current_pages -
|
H A D | pvr_free_list.h | 75 /** @max_pages: Maximum number of pages in free list. */ 76 u32 max_pages; member in struct:pvr_free_list
|
H A D | pvr_rogue_fwif.h | 2061 u32 max_pages; member in struct:rogue_fwif_freelist
|
H A D | pvr_rogue_fwif_check.h | 431 OFFSET_CHECK(struct rogue_fwif_freelist, max_pages, 20);
|
/linux-master/drivers/infiniband/core/ |
H A D | rw.c | 63 u32 max_pages; local 66 max_pages = dev->attrs.max_pi_fast_reg_page_list_len; 68 max_pages = dev->attrs.max_fast_reg_page_list_len; 71 return min_t(u32, max_pages, 256);
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | mlx4_ib.h | 139 u32 max_pages; member in struct:mlx4_ib_mr
|
H A D | mr.c | 551 int max_pages) 557 * max_pages is limited to MLX4_MAX_FAST_REG_PAGES 560 mr->page_map_size = roundup(max_pages * sizeof(u64), 668 mr->max_pages = max_num_sg; 692 if (unlikely(mr->npages == mr->max_pages)) 549 mlx4_alloc_priv_pages(struct ib_device *device, struct mlx4_ib_mr *mr, int max_pages) argument
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | umr.c | 733 size_t max_pages = ib_umem_odp_num_pages(odp) - idx; local 735 pages_to_map = min_t(size_t, pages_to_map, max_pages);
|
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma.h | 147 u32 max_pages; member in struct:pvrdma_user_mr
|
H A D | pvrdma_mr.c | 252 mr->max_pages = max_num_sg; 306 if (mr->npages == mr->max_pages)
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_loc.h | 64 int rxe_mr_init_fast(int max_pages, struct rxe_mr *mr);
|
H A D | rxe_mr.c | 188 int rxe_mr_init_fast(int max_pages, struct rxe_mr *mr) argument 195 err = rxe_mr_alloc(mr, max_pages);
|
/linux-master/drivers/iommu/iommufd/ |
H A D | pages.c | 315 static int __batch_init(struct pfn_batch *batch, size_t max_pages, void *backup, argument 319 size_t size = max_pages * elmsz; 332 static int batch_init(struct pfn_batch *batch, size_t max_pages) argument 334 return __batch_init(batch, max_pages, NULL, 0); 337 static void batch_init_backup(struct pfn_batch *batch, size_t max_pages, argument 340 __batch_init(batch, max_pages, backup, backup_len);
|
/linux-master/drivers/md/bcache/ |
H A D | bcache.h | 772 unsigned int n, max_pages; local 774 max_pages = min_t(unsigned int, 779 if (n > max_pages) 780 n = max_pages;
|
/linux-master/drivers/md/ |
H A D | dm-writecache.c | 1835 unsigned int max_pages; local 1842 max_pages = e->wc_list_contiguous; 1844 bio = bio_alloc_bioset(wc->dev->bdev, max_pages, REQ_OP_WRITE, 1851 if (unlikely(max_pages > WB_LIST_INLINE)) 1852 wb->wc_list = kmalloc_array(max_pages, sizeof(struct wc_entry *), 1856 if (likely(max_pages <= WB_LIST_INLINE) || unlikely(!wb->wc_list)) { 1858 max_pages = WB_LIST_INLINE; 1866 while (wbl->size && wb->wc_list_n < max_pages) {
|
/linux-master/drivers/net/ethernet/emulex/benet/ |
H A D | be_cmds.c | 820 static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages, argument 823 int i, buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages);
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_wq.c | 440 * @max_pages: maximum pages allowed 445 int max_pages) 451 if (num_q_pages > max_pages) { 444 alloc_wq_pages(struct hinic_wq *wq, struct hinic_hwif *hwif, int max_pages) argument
|