/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_chain.c | 12 u32 page_cnt) 34 chain->page_cnt = page_cnt; 35 chain->capacity = chain->usable_per_page * page_cnt; 36 chain->size = chain->elem_per_page * page_cnt; 80 for (i = 0; i < chain->page_cnt; i++) { 114 for (i = 0; i < chain->page_cnt; i++) { 160 u32 page_cnt) 165 chain_size *= page_cnt; 206 for (i = 0; i < chain->page_cnt; 10 qed_chain_init(struct qed_chain *chain, const struct qed_chain_init_params *params, u32 page_cnt) argument 158 qed_chain_alloc_sanity_check(struct qed_dev *cdev, const struct qed_chain_init_params *params, u32 page_cnt) argument 255 u32 page_cnt, i; local 323 u32 page_cnt; local [all...] |
H A D | qed_sp_commands.c | 311 u8 page_cnt, i; local 372 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); 373 p_ramrod->event_ring_num_pages = page_cnt; 378 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_consq->chain); 379 p_ramrod->consolid_q_num_pages = page_cnt;
|
/linux-master/tools/testing/selftests/bpf/ |
H A D | bpf_arena_common.h | 45 void __arena* bpf_arena_alloc_pages(void *map, void __arena *addr, __u32 page_cnt, 47 void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak; 61 static inline void __arena* bpf_arena_alloc_pages(void *map, void *addr, __u32 page_cnt, argument 66 static inline void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) argument
|
/linux-master/kernel/bpf/ |
H A D | arena.c | 406 static long arena_alloc_pages(struct bpf_arena *arena, long uaddr, long page_cnt, int node_id) argument 416 if (page_cnt > page_cnt_max) 423 if (pgoff > page_cnt_max - page_cnt) 429 pages = kvcalloc(page_cnt, sizeof(struct page *), GFP_KERNEL); 436 ret = mtree_insert_range(&arena->mt, pgoff, pgoff + page_cnt - 1, 440 page_cnt, 0, page_cnt_max - 1, GFP_KERNEL); 445 node_id, page_cnt, pages); 450 /* Earlier checks made sure that uaddr32 + page_cnt * PAGE_SIZE - 1 454 * kern_vm_start + uaddr32 + page_cnt * PAGE_SIZE - 1 can overflow 458 kern_vm_start + uaddr32 + page_cnt * PAGE_SIZ 478 zap_pages(struct bpf_arena *arena, long uaddr, long page_cnt) argument 487 arena_free_pages(struct bpf_arena *arena, long uaddr, long page_cnt) argument 532 bpf_arena_alloc_pages(void *p__map, void *addr__ign, u32 page_cnt, int node_id, u64 flags) argument 544 bpf_arena_free_pages(void *p__map, void *ptr__ign, u32 page_cnt) argument [all...] |
/linux-master/include/linux/qed/ |
H A D | qed_chain.h | 106 u32 page_cnt; member in struct:qed_chain 262 return chain->page_cnt; 303 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) 307 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) 493 /* Use (page_cnt - 1) as a reset value for the prod/cons page's 496 * will be advanced to page_cnt and then will be wrapped to 0. 498 u32 reset_val = p_chain->page_cnt - 1; 556 last_page_idx = p_chain->page_cnt - 1; 580 u32 cur_prod, page_mask, page_cnt, page_diff; local 599 page_cnt 626 u32 i, page_cnt; local [all...] |
/linux-master/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-helper-util.c | 133 union cvmx_ipd_portx_bp_page_cnt page_cnt; local 141 page_cnt.u64 = 0; 142 page_cnt.s.bp_enb = 0; 143 page_cnt.s.page_cnt = 100; 148 page_cnt.u64);
|
/linux-master/drivers/base/firmware_loader/ |
H A D | sysfs.c | 246 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); local 250 page_ofs, page_cnt); 253 buffer, page_cnt); 255 buffer += page_cnt; 256 offset += page_cnt; 257 count -= page_cnt;
|
/linux-master/drivers/hv/ |
H A D | ring_buffer.c | 186 struct page *pages, u32 page_cnt, u32 max_pkt_size) 197 pages_wraparound = kcalloc(page_cnt * 2 - 1, 204 for (i = 0; i < 2 * (page_cnt - 1); i++) 206 &pages[i % (page_cnt - 1) + 1]; 209 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, 228 ring_info->ring_size = page_cnt << PAGE_SHIFT; 185 hv_ringbuffer_init(struct hv_ring_buffer_info *ring_info, struct page *pages, u32 page_cnt, u32 max_pkt_size) argument
|
H A D | hv_balloon.c | 146 __u64 page_cnt:24; member in struct:dm_mem_page_range::__anon66 1008 pfn_cnt = dm->ha_wrk.ha_page_range.finfo.page_cnt; 1011 rg_sz = dm->ha_wrk.ha_region_range.finfo.page_cnt; 1220 int num_pages = range_array->finfo.page_cnt; 1279 bl_resp->range_array[i].finfo.page_cnt = alloc_unit;
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 721 int page_cnt = 0; local 726 page_cnt += region->count; 729 return page_cnt; 793 dma_addr_t *pages, unsigned int page_cnt) 810 mapped_cnt < page_cnt; i++) { 818 if (r->offset + r->count > page_cnt) { 822 i, r->offset, r->count, page_cnt); 827 page_cnt - mapped_cnt); 838 if (mapped_cnt < page_cnt) { 841 mapped_cnt, page_cnt); 792 hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr, dma_addr_t *pages, unsigned int page_cnt) argument 1017 int page_cnt, i; local [all...] |
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | encl.h | 55 unsigned int page_cnt; member in struct:sgx_encl
|
H A D | ioctl.c | 28 if (!(encl->page_cnt % SGX_VA_SLOT_COUNT)) { 40 WARN_ON_ONCE(encl->page_cnt % SGX_VA_SLOT_COUNT); 42 encl->page_cnt++; 48 encl->page_cnt--;
|
/linux-master/arch/mips/include/asm/octeon/ |
H A D | cvmx-ipd-defs.h | 324 uint64_t page_cnt:17; member in struct:cvmx_ipd_bpidx_mbuf_th::cvmx_ipd_bpidx_mbuf_th_s 326 uint64_t page_cnt:17; 999 uint64_t page_cnt:17; member in struct:cvmx_ipd_portx_bp_page_cnt::cvmx_ipd_portx_bp_page_cnt_s 1001 uint64_t page_cnt:17; 1014 uint64_t page_cnt:17; member in struct:cvmx_ipd_portx_bp_page_cnt2::cvmx_ipd_portx_bp_page_cnt2_s 1016 uint64_t page_cnt:17; 1029 uint64_t page_cnt:17; member in struct:cvmx_ipd_portx_bp_page_cnt3::cvmx_ipd_portx_bp_page_cnt3_s 1031 uint64_t page_cnt:17; 1387 uint64_t page_cnt:25; member in struct:cvmx_ipd_sub_port_bp_page_cnt::cvmx_ipd_sub_port_bp_page_cnt_s 1389 uint64_t page_cnt [all...] |
/linux-master/drivers/staging/rts5208/ |
H A D | xd.c | 1499 u8 reg_val, page_cnt; local 1505 page_cnt = end_page - start_page; 1531 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); 1536 page_cnt * 512, DMA_512); 1546 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, 1701 u8 page_cnt, reg_val; local 1709 page_cnt = end_page - start_page; 1731 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); 1735 page_cnt * 512, DMA_512); 1744 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 51 1831 u8 start_page, end_page = 0, page_cnt; local [all...] |
/linux-master/fs/ntfs3/ |
H A D | fslog.c | 1584 u32 page_cnt = 1, page_pos = 1; local 1707 page_cnt = le16_to_cpu(best_page->page_count); 1708 if (page_cnt > 1) 1716 (page_pos != page_cnt || best_page_pos == page_pos || 1718 (page_pos >= page_cnt || best_page_pos == page_pos)) { 1730 page_cnt = bppc; 1733 if (page_cnt > 1) 1736 page_pos = page_cnt = 1; 1766 page_cnt = page_pos = 1; 1808 ((page_pos == page_cnt [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | verbs.h | 106 u32 page_cnt; member in struct:irdma_mr
|
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 142 mr->mem.page_cnt > ERDMA_MAX_INLINE_MTT_ENTRIES) { 152 MTT_SIZE(mr->mem.page_cnt)); 164 FIELD_PREP(ERDMA_CMD_REGMR_MTT_CNT_MASK, mr->mem.page_cnt); 179 req.mtt_cnt_h = mr->mem.page_cnt >> 20; 772 mem->page_cnt = mem->mtt_nents; 773 mem->mtt = erdma_create_mtt(dev, MTT_SIZE(mem->page_cnt), 1091 mr->mem.page_cnt = max_num_sg; 1121 if (mr->mem.mtt_nents >= mr->mem.page_cnt)
|
H A D | erdma_verbs.h | 117 u32 page_cnt; member in struct:erdma_mem
|
/linux-master/drivers/infiniband/hw/efa/ |
H A D | efa_verbs.c | 1243 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) argument 1249 sglist = kmalloc_array(page_cnt, sizeof(*sglist), GFP_KERNEL); 1252 sg_init_table(sglist, page_cnt); 1253 for (i = 0; i < page_cnt; i++) { 1274 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; local 1285 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); 1296 page_cnt); 1307 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) +
|
/linux-master/fs/ubifs/ |
H A D | file.c | 718 int err, page_idx, page_cnt, ret = 0, n = 0; local 733 page_cnt = bu->blk_cnt >> UBIFS_BLOCKS_PER_PAGE_SHIFT; 734 if (!page_cnt) { 777 for (page_idx = 1; page_idx < page_cnt; page_idx++) {
|
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | main.c | 788 u32 page_cnt; local 802 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); 803 cur_pbl->num_pbl_pages = page_cnt;
|
/linux-master/drivers/net/wireless/realtek/rtw88/ |
H A D | fw.c | 2002 u8 page_cnt, pages; local 2007 page_cnt = RTW_OLD_PROBE_PG_CNT; 2009 page_cnt = RTW_PROBE_PG_CNT; 2011 pages = page_offset + num_probes * page_cnt; 2021 if (skb->len > page_size * page_cnt) { 2031 buf_offset += page_cnt * page_size; 2032 page_offset += page_cnt;
|
/linux-master/drivers/target/ |
H A D | target_core_user.c | 513 int page_cnt = DIV_ROUND_UP(length, PAGE_SIZE); local 523 for (cnt = 0; xas_next(&xas) && cnt < page_cnt;) 527 for (i = cnt; i < page_cnt; i++) { 551 return i == page_cnt ? dbi : -1; 710 int page_cnt, page_inx, dpi; local 729 page_cnt = DIV_ROUND_UP(data_len, PAGE_SIZE); 730 if (page_cnt > udev->data_pages_per_blk) 731 page_cnt = udev->data_pages_per_blk; 734 for (page_inx = 0; page_inx < page_cnt && data_len;
|
/linux-master/tools/lib/bpf/ |
H A D | libbpf.h | 1491 * @param page_cnt number of memory pages allocated for each per-CPU buffer 1499 perf_buffer__new(int map_fd, size_t page_cnt, 1533 perf_buffer__new_raw(int map_fd, size_t page_cnt, struct perf_event_attr *attr,
|
/linux-master/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_main.c | 2193 u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl); local 2214 page_cnt, &ret_params); 2283 u32 page_cnt; local 2299 page_cnt = qed_chain_get_page_cnt(&rxq->rx_comp_ring); 2305 page_cnt, &ret_params);
|