/linux-master/tools/perf/util/bpf_skel/ |
H A D | lock_contention.bpf.c | 27 __uint(max_entries, MAX_ENTRIES); 35 __uint(max_entries, MAX_ENTRIES); 43 __uint(max_entries, 1); 51 __uint(max_entries, MAX_ENTRIES); 58 __uint(max_entries, MAX_ENTRIES); 65 __uint(max_entries, MAX_ENTRIES); 72 __uint(max_entries, 1); 79 __uint(max_entries, 1); 86 __uint(max_entries, 1); 93 __uint(max_entries, [all...] |
/linux-master/io_uring/ |
H A D | io_uring.c | 1176 * If more entries than max_entries are available, stop processing once this 1181 unsigned int max_entries) 1208 } while (node && *count < max_entries); 1257 unsigned int max_entries, 1270 node = io_handle_tw_list(node, count, max_entries); 1179 io_handle_tw_list(struct llist_node *node, unsigned int *count, unsigned int max_entries) argument 1256 tctx_task_work_run(struct io_uring_task *tctx, unsigned int max_entries, unsigned int *count) argument
|
/linux-master/fs/ceph/ |
H A D | mds_client.c | 2569 req->r_args.readdir.max_entries = cpu_to_le32(num_entries);
|
/linux-master/net/xdp/ |
H A D | xsk.c | 480 static u32 xsk_tx_peek_release_fallback(struct xsk_buff_pool *pool, u32 max_entries) argument 485 while (nb_pkts < max_entries && xsk_tx_peek_desc(pool, &descs[nb_pkts]))
|
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 6047 u16 max_rings, max_entries, pad, i; local 6057 max_entries = bnxt_get_rxfh_indir_size(bp->dev); 6059 for (i = 0; i < max_entries; i++) 6062 pad = bp->rss_indir_tbl_entries - max_entries; 7986 if (!ctxm->max_entries) 8041 ctxm->max_entries = le32_to_cpu(resp->max_num_entries); 8097 ctxm->max_entries = le32_to_cpu(resp->qp_max_entries); 8107 ctxm->max_entries = le32_to_cpu(resp->srq_max_entries); 8114 ctxm->max_entries = le32_to_cpu(resp->cq_max_entries); 8121 ctxm->max_entries [all...] |
/linux-master/net/core/ |
H A D | sock_map.c | 35 if (attr->max_entries == 0 || 49 stab->sks = bpf_map_area_alloc((u64) stab->map.max_entries * 342 for (i = 0; i < stab->map.max_entries; i++) { 376 if (unlikely(key >= map->max_entries)) 445 if (unlikely(i >= map->max_entries)) 458 if (i == stab->map.max_entries - 1) 460 if (i >= stab->map.max_entries) 479 if (unlikely(idx >= map->max_entries)) 713 if (unlikely(info->index >= info->map->max_entries)) 807 usage += (u64)map->max_entries * sizeo [all...] |
/linux-master/kernel/trace/ |
H A D | bpf_trace.c | 561 if (unlikely(index >= array->map.max_entries)) 633 if (unlikely(index >= array->map.max_entries)) 806 if (unlikely(idx >= array->map.max_entries))
|
/linux-master/kernel/bpf/ |
H A D | verifier.c | 10048 max = map->max_entries; 11300 * So, in case of global variables, they use array maps with max_entries = 1, 11302 * into the same map value as max_entries is 1, as described above). 15251 WARN_ON_ONCE(map->max_entries != 1); 19846 * if (index >= max_entries) goto out; 19857 map_ptr->max_entries, 2);
|
H A D | syscall.c | 344 map->max_entries = attr->max_entries; 853 "max_entries:\t%u\n" 862 map->max_entries, 4847 info.max_entries = map->max_entries;
|
H A D | bloom_filter.c | 102 attr->max_entries == 0 || 127 if (check_mul_overflow(attr->max_entries, nr_hash_funcs, &nr_bits) ||
|
H A D | arena.c | 101 if (attr->key_size || attr->value_size || attr->max_entries == 0 || 112 vm_range = (u64)attr->max_entries * PAGE_SIZE;
|
/linux-master/include/linux/ |
H A D | bpf.h | 266 u32 max_entries; member in struct:bpf_map
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_npc.c | 1497 int entry, num_entries, max_entries; local 1506 max_entries = rvu->hw->npc_kpu_entries; 1509 num_entries = min_t(int, profile->cam_entries, max_entries); 1515 num_entries = min_t(int, profile->action_entries, max_entries);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_prime.c | 991 * @max_entries: size of the passed-in array 1001 int max_entries) 1007 if (WARN_ON(p - pages >= max_entries)) 1019 * @max_entries: size of both the passed-in arrays 1027 int max_entries) 1033 if (WARN_ON(a - addrs >= max_entries)) 999 drm_prime_sg_to_page_array(struct sg_table *sgt, struct page **pages, int max_entries) argument 1026 drm_prime_sg_to_dma_addr_array(struct sg_table *sgt, dma_addr_t *addrs, int max_entries) argument
|
/linux-master/arch/x86/net/ |
H A D | bpf_jit_comp.c | 596 * if (index >= array->map.max_entries) 622 * if (index >= array->map.max_entries) 627 offsetof(struct bpf_array, map.max_entries));
|
/linux-master/include/ufs/ |
H A D | ufshcd.h | 1100 * @max_entries: max number of slots in this hardware queue 1119 u32 max_entries; member in struct:ufs_hw_queue
|
/linux-master/include/uapi/scsi/ |
H A D | scsi_bsg_mpi3mr.h | 216 * @max_entries: Number of log data entries cached by the driver 220 __u16 max_entries; member in struct:mpi3mr_logdata_enable
|
/linux-master/drivers/ufs/core/ |
H A D | ufs-mcq.c | 209 hwq->max_entries; 218 cqe_size = sizeof(struct cq_entry) * hwq->max_entries; 292 u32 entries = hwq->max_entries; 337 qsize = hwq->max_entries * MCQ_ENTRY_SIZE_IN_DWORD - 1; 443 hwq->max_entries = hba->nutrs + 1; 614 if (sq_head_slot == hwq->max_entries)
|
/linux-master/drivers/scsi/mpi3mr/ |
H A D | mpi3mr_app.c | 157 logdata_enable.max_entries =
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | verifier_arena.c | 14 __uint(max_entries, 2); /* arena of two pages close to 32-bit boundary*/ 110 pages = bpf_arena_alloc_pages(&ar->map, NULL, ar->map.max_entries, NUMA_NO_NODE, 0); 124 bpf_arena_alloc_pages(map, NULL, map->max_entries, 0, 0); 146 bpf_arena_alloc_pages(map->inner_map_meta, NULL, map->max_entries, 0, 0);
|
H A D | verifier_arena_large.c | 16 __uint(max_entries, ARENA_SIZE / PAGE_SIZE);
|
H A D | arena_list.c | 12 __uint(max_entries, 100); /* number of pages */
|
H A D | arena_htab.c | 12 __uint(max_entries, 100); /* number of pages */
|
/linux-master/tools/lib/bpf/ |
H A D | libbpf.c | 515 unsigned int max_entries; member in struct:bpf_map_def 1345 map->def.max_entries = 1; 1700 static size_t array_map_mmap_sz(unsigned int value_sz, unsigned int max_entries) argument 1705 map_sz = (size_t)roundup(value_sz, 8) * max_entries; 1716 return array_map_mmap_sz(map->def.value_size, map->def.max_entries); 1718 return page_sz * map->def.max_entries; 1864 def->max_entries = 1; 2462 } else if (strcmp(name, "max_entries") == 0) { 2463 if (!get_map_field_int(map_name, btf, m, &map_def->max_entries)) 2669 /* Kernel expects BPF_MAP_TYPE_RINGBUF's max_entries t 4823 bpf_map__set_max_entries(struct bpf_map *map, __u32 max_entries) argument [all...] |
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_main.c | 3121 int max_entries = hw->mac.rar_entry_count; local 3125 for (i = 0; i < max_entries; i++) { 3148 int max_entries = hw->mac.rar_entry_count; local 3152 for (i = 0; i < max_entries; i++) {
|