/linux-master/drivers/virt/vboxguest/ |
H A D | vboxguest_core.h | 43 /** The current number of chunks in the balloon. */ 44 u32 chunks; member in struct:vbg_mem_balloon 45 /** The maximum number of chunks in the balloon. */
|
/linux-master/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-srv.h | 111 struct page **chunks; member in struct:rtrs_srv_sess
|
H A D | rtrs-srv.c | 577 * Here we map queue_depth chunks to MR. Firstly we have to 578 * figure out how many chunks can we map per MR. 582 * in order to do invalidate for each chunks of memory, we needs 601 int nr, nr_sgt, chunks; local 604 chunks = chunks_per_mr * srv_path->mrs_num; 607 srv->queue_depth - chunks); 614 sg_set_page(s, srv->chunks[chunks + i], 649 srv_path->dma_addr[chunks + i] = sg_dma_address(s); 1035 data = page_address(srv->chunks[buf_i [all...] |
/linux-master/drivers/gpu/drm/sprd/ |
H A D | sprd_dsi.c | 460 u32 chunks = 0; local 539 chunks = vm->hactive / video_size; 542 if (total_bytes >= (bytes_per_chunk * chunks)) { 544 bytes_per_chunk * chunks; 550 if (bytes_left > (pkt_header * chunks)) { 552 pkt_header * chunks) / chunks; 560 chunks = 1; 571 dsi_reg_wr(ctx, VIDEO_PKT_CONFIG, VIDEO_LINE_CHUNK_NUM, 16, chunks);
|
/linux-master/arch/x86/kernel/cpu/resctrl/ |
H A D | monitor.c | 265 u64 shift = 64 - width, chunks; local 267 chunks = (cur_msr << shift) - (prev_msr << shift); 268 return chunks >> shift; 278 u64 msr_val, chunks; local 292 am->chunks += mbm_overflow_count(am->prev_msr, msr_val, 294 chunks = get_corrected_mbm_count(rmid, am->chunks); 297 chunks = msr_val; 300 *val = chunks * hw_res->mon_scale; 569 * and delta bandwidth in MBps. The chunks valu [all...] |
/linux-master/net/sctp/ |
H A D | auth.c | 186 struct sctp_chunks_param *chunks, 197 if (chunks) 198 chunks_len = ntohs(chunks->param_hdr.length); 209 if (chunks) { 210 memcpy(new->data + offset, chunks, chunks_len); 651 * chunks MUST NOT be listed in the CHUNKS parameter. However, if 653 * SHUTDOWN-COMPLETE and AUTH chunks MUST be ignored. 656 switch (param->chunks[i]) { 664 if (param->chunks[i] == chunk) 704 * zero (as shown in Figure 6) followed by all chunks tha 184 sctp_auth_make_key_vector( struct sctp_random_param *random, struct sctp_chunks_param *chunks, struct sctp_hmac_algo_param *hmacs, gfp_t gfp) argument [all...] |
H A D | stream_sched_rr.c | 104 ch = list_first_entry(&msg->chunks, struct sctp_chunk, frag_list);
|
H A D | stream_sched_prio.c | 222 ch = list_first_entry(&msg->chunks, struct sctp_chunk, frag_list);
|
H A D | stream_sched_fc.c | 105 ch = list_first_entry(&msg->chunks, struct sctp_chunk, frag_list);
|
H A D | stream_sched.c | 178 /* We have to requeue all chunks already queued. */ 237 if (!list_is_last(&ch->frag_list, &ch->msg->chunks) &&
|
/linux-master/drivers/dma/sh/ |
H A D | shdma-base.c | 85 /* Mark all chunks of this descriptor as submitted, move to the queue */ 88 * All chunks are on the global ld_free, so, we have to find 97 if (chunk->chunks == 1) { 134 * Make it int then, on error remove chunks from the 347 * any chunks in a (partially) completed chain 356 if (desc->mark == DESC_COMPLETED && desc->chunks == 1) { 372 BUG_ON(desc->chunks != 1); 567 int chunks = 0; local 572 chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len); 582 * if more than one chunk is needed further chunks hav [all...] |
H A D | rcar-dmac.c | 29 * @node: entry in the parent's chunks list 62 * @chunks: list of transfer chunks for this transfer 64 * @nchunks: number of transfer chunks for this transfer 79 struct list_head chunks; member in struct:rcar_dmac_desc 100 * @chunks: array of transfer chunk descriptors 107 DECLARE_FLEX_ARRAY(struct rcar_dmac_xfer_chunk, chunks); 115 ((PAGE_SIZE - offsetof(struct rcar_dmac_desc_page, chunks)) / \ 389 list_first_entry(&desc->chunks, 513 desc->running = list_first_entry(&desc->chunks, [all...] |
/linux-master/drivers/mtd/nand/raw/ |
H A D | davinci_nand.c | 536 int chunks = mtd->writesize / 512; local 538 if (!chunks || mtd->oobsize < 16) { 576 if (chunks == 1) { 579 } else if (chunks == 4 || chunks == 8) {
|
/linux-master/include/linux/ |
H A D | genalloc.h | 60 struct list_head chunks; /* list of chunks in this pool */ member in struct:gen_pool
|
H A D | sctp.h | 321 __u8 chunks[]; member in struct:sctp_supported_ext_param 333 __u8 chunks[]; member in struct:sctp_chunks_param 369 * chunks and to inform the peer endpoint of gaps in the received 370 * subsequences of DATA chunks as represented by their TSNs. 597 * chunks this field MUST be filled in. 696 * This value MUST be set to 0x0F for all AUTH-chunks.
|
/linux-master/drivers/md/ |
H A D | md-bitmap.h | 191 unsigned long chunks; /* Total number of data member in struct:bitmap::bitmap_counts 192 * chunks for the array */
|
/linux-master/net/xdp/ |
H A D | xsk_buff_pool.c | 63 entries = unaligned ? umem->chunks : 0; 68 pool->heads = kvcalloc(umem->chunks, sizeof(*pool->heads), GFP_KERNEL); 78 pool->heads_cnt = umem->chunks; 79 pool->free_heads_cnt = umem->chunks;
|
/linux-master/include/net/ |
H A D | xdp_sock.h | 28 u32 chunks; member in struct:xdp_umem
|
/linux-master/drivers/net/wireless/intel/iwlwifi/fw/ |
H A D | pnvm.c | 126 pnvm_data->chunks[pnvm_data->n_chunks].data = section->data; 127 pnvm_data->chunks[pnvm_data->n_chunks].len = data_len;
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 83 struct list_head chunks; member in struct:nouveau_dmem 275 list_add(&chunk->list, &drm->dmem->chunks); 348 list_for_each_entry(chunk, &drm->dmem->chunks, list) { 365 list_for_each_entry(chunk, &drm->dmem->chunks, list) 426 list_for_each_entry_safe(chunk, tmp, &drm->dmem->chunks, list) { 603 INIT_LIST_HEAD(&drm->dmem->chunks);
|
/linux-master/drivers/dma/ioat/ |
H A D | dma.c | 378 int i, chunks; local 385 chunks = (total_descs * IOAT_DESC_SZ) / IOAT_CHUNK_SIZE; 386 ioat_chan->desc_chunks = chunks; 388 for (i = 0; i < chunks; i++) { 448 if (chunks == 1)
|
/linux-master/lib/ |
H A D | scatterlist.c | 276 * defined to return scatterlist chunks of maximum size @max_ents. 459 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; local 499 /* compute number of contiguous chunks */ 500 chunks = 1; 506 chunks++; 511 /* merging chunks and putting them into the scatterlist */ 513 for (i = 0; i < chunks; i++) { 525 /* Pass how many chunks might be left */ 526 s = get_next_sg(sgt_append, s, chunks - i + left_pages,
|
H A D | debugobjects.c | 961 unsigned long flags, oaddr, saddr, eaddr, paddr, chunks; local 970 chunks = ((eaddr - paddr) + (ODEBUG_CHUNK_SIZE - 1)); 971 chunks >>= ODEBUG_CHUNK_SHIFT; 973 for (;chunks > 0; chunks--, paddr += ODEBUG_CHUNK_SIZE) {
|
/linux-master/fs/xfs/ |
H A D | xfs_buf_item.c | 82 * stretch of non-contiguous chunks to be logged. Contiguous chunks are logged 113 * chunks of unmapped buffers anywhere. 318 * Fill in an iovec for each set of contiguous chunks. 328 * chunks of unmapped buffers anywhere. 393 * format structure, and the rest point to contiguous chunks 833 int chunks; local 855 * chunks is the number of XFS_BLF_CHUNK size pieces the buffer 858 * chunks of the buffer. 866 chunks [all...] |
/linux-master/fs/smb/server/ |
H A D | vfs.h | 97 struct srv_copychunk *chunks,
|