Searched refs:max_pages (Results 1 - 25 of 62) sorted by path

123

/linux-master/arch/x86/xen/
H A Dsetup.c513 unsigned long max_pages, limit; local
518 max_pages = limit;
532 max_pages = ret;
535 return min(max_pages, limit);
692 unsigned long max_pages; local
750 max_pages = xen_get_max_pages();
753 max_pages += xen_foreach_remap_area(max_pfn, xen_count_remap_pages);
755 if (max_pages > max_pfn)
756 extra_pages += max_pages - max_pfn;
762 * Make sure we have no memory above max_pages, a
[all...]
/linux-master/crypto/
H A Dalgif_hash.c70 size_t len, max_pages, npages; local
74 max_pages = min_t(size_t, ALG_MAX_PAGES,
111 npages = iov_iter_npages(&msg->msg_iter, max_pages);
/linux-master/drivers/base/
H A Dmemory.c1103 * @max_pages: The maximum number of pages we'll have in this static memory
1112 * memory groups can be registered, or if max_pages is invalid (0). Otherwise,
1115 int memory_group_register_static(int nid, unsigned long max_pages) argument
1120 .max_pages = max_pages,
1124 if (!max_pages)
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_svm.c1989 uint64_t max_pages; local
2006 max_pages = READ_ONCE(max_svm_range_pages);
2007 _pages = min_not_zero(max_pages, min_pages);
2008 } while (cmpxchg(&max_svm_range_pages, max_pages, _pages) != max_pages);
2013 uint64_t max_pages, struct list_head *insert_list,
2020 max_pages, start, last);
2023 l = min(last, ALIGN_DOWN(start + max_pages, max_pages) - 1);
2012 svm_range_split_new(struct svm_range_list *svms, uint64_t start, uint64_t last, uint64_t max_pages, struct list_head *insert_list, struct list_head *update_list) argument
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_pages.c719 unsigned long max_pages; local
740 max_pages = vm->total >> PAGE_SHIFT;
742 for_each_prime_number_from(page_num, 1, max_pages) {
/linux-master/drivers/gpu/drm/i915/selftests/
H A Di915_gem_gtt.c416 const unsigned long max_pages = local
418 const unsigned long max_step = max(int_sqrt(max_pages), 2UL);
431 for (npages = 1; npages <= max_pages; npages *= prime) {
638 const unsigned long max_pages = local
652 for_each_prime_number_from(size, 1, max_pages) {
H A Di915_vma.c577 const unsigned int max_pages = 64; local
584 obj = i915_gem_object_create_internal(vm->i915, max_pages * PAGE_SIZE);
601 GEM_BUG_ON(max_offset > max_pages);
602 max_offset = max_pages - max_offset;
H A Dintel_memory_region.c60 unsigned long max_pages; local
66 max_pages = div64_u64(total, page_size);
69 for_each_prime_number_from(page_num, 1, max_pages) {
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_free_list.c91 free_list->max_pages = args->max_num_pages;
179 fw_data->max_pages = free_list->max_pages;
188 ((fw_data->max_pages - fw_data->current_pages) *
280 start_page = free_list->max_pages - free_list->current_pages -
H A Dpvr_free_list.h75 /** @max_pages: Maximum number of pages in free list. */
76 u32 max_pages; member in struct:pvr_free_list
H A Dpvr_rogue_fwif.h2061 u32 max_pages; member in struct:rogue_fwif_freelist
H A Dpvr_rogue_fwif_check.h431 OFFSET_CHECK(struct rogue_fwif_freelist, max_pages, 20);
/linux-master/drivers/infiniband/core/
H A Drw.c63 u32 max_pages; local
66 max_pages = dev->attrs.max_pi_fast_reg_page_list_len;
68 max_pages = dev->attrs.max_fast_reg_page_list_len;
71 return min_t(u32, max_pages, 256);
/linux-master/drivers/infiniband/hw/mlx4/
H A Dmlx4_ib.h139 u32 max_pages; member in struct:mlx4_ib_mr
H A Dmr.c551 int max_pages)
557 * max_pages is limited to MLX4_MAX_FAST_REG_PAGES
560 mr->page_map_size = roundup(max_pages * sizeof(u64),
668 mr->max_pages = max_num_sg;
692 if (unlikely(mr->npages == mr->max_pages))
549 mlx4_alloc_priv_pages(struct ib_device *device, struct mlx4_ib_mr *mr, int max_pages) argument
/linux-master/drivers/infiniband/hw/mlx5/
H A Dumr.c733 size_t max_pages = ib_umem_odp_num_pages(odp) - idx; local
735 pages_to_map = min_t(size_t, pages_to_map, max_pages);
/linux-master/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma.h147 u32 max_pages; member in struct:pvrdma_user_mr
H A Dpvrdma_mr.c252 mr->max_pages = max_num_sg;
306 if (mr->npages == mr->max_pages)
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_loc.h64 int rxe_mr_init_fast(int max_pages, struct rxe_mr *mr);
H A Drxe_mr.c188 int rxe_mr_init_fast(int max_pages, struct rxe_mr *mr) argument
195 err = rxe_mr_alloc(mr, max_pages);
/linux-master/drivers/iommu/iommufd/
H A Dpages.c315 static int __batch_init(struct pfn_batch *batch, size_t max_pages, void *backup, argument
319 size_t size = max_pages * elmsz;
332 static int batch_init(struct pfn_batch *batch, size_t max_pages) argument
334 return __batch_init(batch, max_pages, NULL, 0);
337 static void batch_init_backup(struct pfn_batch *batch, size_t max_pages, argument
340 __batch_init(batch, max_pages, backup, backup_len);
/linux-master/drivers/md/bcache/
H A Dbcache.h772 unsigned int n, max_pages; local
774 max_pages = min_t(unsigned int,
779 if (n > max_pages)
780 n = max_pages;
/linux-master/drivers/md/
H A Ddm-writecache.c1835 unsigned int max_pages; local
1842 max_pages = e->wc_list_contiguous;
1844 bio = bio_alloc_bioset(wc->dev->bdev, max_pages, REQ_OP_WRITE,
1851 if (unlikely(max_pages > WB_LIST_INLINE))
1852 wb->wc_list = kmalloc_array(max_pages, sizeof(struct wc_entry *),
1856 if (likely(max_pages <= WB_LIST_INLINE) || unlikely(!wb->wc_list)) {
1858 max_pages = WB_LIST_INLINE;
1866 while (wbl->size && wb->wc_list_n < max_pages) {
/linux-master/drivers/net/ethernet/emulex/benet/
H A Dbe_cmds.c820 static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages, argument
823 int i, buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages);
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_wq.c440 * @max_pages: maximum pages allowed
445 int max_pages)
451 if (num_q_pages > max_pages) {
444 alloc_wq_pages(struct hinic_wq *wq, struct hinic_hwif *hwif, int max_pages) argument

Completed in 424 milliseconds

123