/linux-master/net/xfrm/ |
H A D | espintcp.c | 262 struct espintcp_msg *emsg = &ctx->partial; 291 struct espintcp_msg *emsg = &ctx->partial; 326 struct espintcp_msg *emsg = &ctx->partial; 530 struct espintcp_msg *emsg = &ctx->partial;
|
/linux-master/fs/sysv/ |
H A D | itree.c | 214 Indirect *partial; local 222 partial = get_branch(inode, depth, offsets, chain, &err); 225 if (!partial) { 230 partial = chain+depth-1; /* the whole chain */ 237 while (partial > chain) { 238 brelse(partial->bh); 239 partial--; 253 left = (chain + depth) - partial; 254 err = alloc_branch(inode, left, offsets+(partial-chain), partial); 286 Indirect *partial, *p; local 372 Indirect *partial; local [all...] |
/linux-master/fs/erofs/ |
H A D | zdata.c | 82 /* L: whether partial decompression or not */ 83 bool partial; member in struct:z_erofs_pcluster 509 * | tail (partial) page | head (partial) page | 787 pcl->partial = true; 1009 * Ensure the current partial folio belongs to this submit chain rather 1037 fe->pcl->partial = false; 1270 .partial_decoding = pcl->partial, 1316 pcl->partial = true; 1739 * Since partial uptodat [all...] |
/linux-master/fs/ |
H A D | buffer.c | 2162 bool partial = false; local 2174 partial = true; 2187 * If this is a partial write which happened to make all buffers 2190 * uptodate as a result of this (potentially partial) write. 2192 if (!partial) 2198 * bringing partial write blocks uptodate first. 2236 * and overwriting a partial write. However if we have 2239 * read_folio might come in and destroy our partial write.
|
H A D | splice.c | 220 buf->offset = spd->partial[page_nr].offset; 221 buf->len = spd->partial[page_nr].len; 222 buf->private = spd->partial[page_nr].private; 269 * Check if we need to grow the arrays holding pages and partial page 281 spd->partial = kmalloc_array(max_usage, sizeof(struct partial_page), 284 if (spd->pages && spd->partial) 288 kfree(spd->partial); 298 kfree(spd->partial); 752 /* dismiss the fully eaten buffers, adjust the partial one */
|
/linux-master/fs/afs/ |
H A D | flock.c | 458 bool partial, no_server_lock = false; local 472 partial = (fl->fl_start != 0 || fl->fl_end != OFFSET_MAX); 474 if (mode == afs_flock_mode_write && partial) 484 * method of upgrade/downgrade, so we need to emulate for partial-file 488 * keeps partial-file locks local. Allow this behaviour to be emulated 492 (partial && mode == afs_flock_mode_openafs)) {
|
/linux-master/fs/ext4/ |
H A D | move_extent.c | 177 int i, err, nr = 0, partial = 0; local 195 partial = 1; 228 if (!partial)
|
H A D | indirect.c | 244 * @partial: pointer to the last triple within a chain 252 Indirect *partial) 260 goal = ext4_find_near(inode, partial); 316 * we had read the existing part of chain and partial points to the last 538 Indirect *partial; local 554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); 557 if (!partial) { 580 * Count number blocks in a subtree under 'partial'. At each 586 for (i = partial - chain + 1; i < depth; i++) 620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); 251 ext4_find_goal(struct inode *inode, ext4_lblk_t block, Indirect *partial) argument 808 Indirect *partial, *p; local 1126 Indirect *partial; local 1242 Indirect *partial, *partial2; local [all...] |
H A D | extents_status.c | 1057 bool partial; member in struct:rsvd_count 1083 * the region to be removed, if any, and note that there's no partial 1096 rc->partial = false; 1111 * Tracks partial clusters found at the beginning and end of extents so 1146 * if we're tracking a partial cluster and the current extent 1149 if (rc->partial && (rc->lclu != EXT4_B2C(sbi, i))) { 1151 rc->partial = false; 1161 rc->partial = false; 1177 * start tracking a partial cluster if there's a partial a [all...] |
/linux-master/tools/perf/scripts/python/ |
H A D | exported-sql-viewer.py | 3557 def __init__(self, glb, title, items, partial, parent=None): 3569 self.partial = partial 3619 if self.partial:
|
/linux-master/drivers/scsi/smartpqi/ |
H A D | smartpqi.h | 284 u8 partial : 1; member in struct:pqi_raid_path_request 309 u8 partial : 1; member in struct:pqi_aio_path_request 340 u8 partial : 1; member in struct:pqi_aio_r1_path_request 373 u8 partial : 1; member in struct:pqi_aio_r56_path_request
|
/linux-master/lib/ |
H A D | test_firmware.c | 62 * @partial: partial read opt when calling request_firmware_into_buf 105 bool partial; member in struct:test_config 235 test_fw_config->partial = false; 313 "partial:\t\t%s\n", 314 test_fw_config->partial ? "true" : "false"); 599 &test_fw_config->partial); 606 return test_dev_config_show_bool(buf, test_fw_config->partial); 873 if (test_fw_config->partial)
|
/linux-master/drivers/gpu/drm/tiny/ |
H A D | repaper.c | 93 bool partial; member in struct:repaper_epd 526 /* repaper can't do partial updates */ 553 if (epd->partial) { 564 epd->partial = true; 579 epd->partial = true; 769 epd->partial = false;
|
/linux-master/fs/ocfs2/ |
H A D | aops.c | 414 int *partial, 431 if (partial && !buffer_uptodate(bh)) 432 *partial = 1; 728 * have to deal with allocating writes to partial pages. 410 walk_page_buffers( handle_t *handle, struct buffer_head *head, unsigned from, unsigned to, int *partial, int (*fn)( handle_t *handle, struct buffer_head *bh)) argument
|
H A D | alloc.c | 5398 * Changing the leftmost offset (via partial or whole 6815 int ret, partial = 0; local 6832 from, to, &partial, 6843 if (!partial) 6933 * Zero partial cluster for a hole punch or truncate. This avoids exposing 7276 * - remove a partial record
|
H A D | refcounttree.c | 2908 int ret = 0, partial; local 2977 from, to, &partial,
|
H A D | move_extents.c | 37 int partial; member in struct:ocfs2_move_extents_context 192 * XXX: defrag can end up with finishing partial extent as requested, 198 int ret, credits = 0, extra_blocks = 0, partial = context->partial; local 287 * allowing partial extent moving is kind of 'pros and cons', it makes 294 if (!partial) { 310 if (partial && (new_len != *len)) 1039 context->partial = 1;
|
/linux-master/drivers/crypto/stm32/ |
H A D | stm32-crc32.c | 69 u32 partial; /* crc32c: partial in first 4 bytes of that struct */ member in struct:stm32_crc_desc_ctx 136 /* Store partial result */ 137 ctx->partial = readl_relaxed(crc->regs + CRC_DR); 163 ctx->partial = crc32_le(ctx->partial, d8, length); 165 ctx->partial = __crc32c_le(ctx->partial, d8, length); 176 writel_relaxed(bitrev32(ctx->partial), crc->regs + CRC_INIT); 205 /* Store partial resul [all...] |
/linux-master/include/linux/ |
H A D | splice.h | 59 struct partial_page *partial; /* pages[] may not be contig */ member in struct:splice_pipe_desc 61 unsigned int nr_pages_max; /* pages[] & partial[] arrays size */
|
/linux-master/arch/x86/crypto/ |
H A D | aesni-intel_avx-x86_64.S | 489 # GCM_COMPLETE Finishes update of tag of last partial block 712 # PARTIAL_BLOCK: Handles encryption/decryption and the tag partial blocks 715 # Outputs encrypted bytes, and updates hash and partial info in gcm_data_context 721 je .L_partial_block_done_\@ # Leave Macro if no partial blocks 756 # Determine if partial block is not being filled and 792 # Determine if partial block is not being filled and
|
/linux-master/tools/power/pm-graph/ |
H A D | sleepgraph.py | 2417 self.partial = False 2565 self.partial = True
|
/linux-master/drivers/dma/qcom/ |
H A D | bam_dma.c | 1082 u32 partial = MAX_DESCRIPTORS - bchan->tail; local 1085 partial * sizeof(struct bam_desc_hw)); 1086 memcpy(fifo, &desc[partial], 1087 (async_desc->xfer_len - partial) *
|
/linux-master/fs/reiserfs/ |
H A D | inode.c | 2515 int partial = 0; local 2669 partial = 1; 2674 if (!partial)
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | puda.c | 1193 * @buf: rcv buffer with partial 1210 * irdma_ieq_setup_tx_buf - setup tx buffer for partial handling 1211 * @buf: reeive buffer with partial 1234 * @fps: first partial sequence number 1294 /* copy partial fpdu */ 1359 * irdma_ieq_handle_partial - process partial fpdu buffer 1361 * @pfpdu: partial management per user qp 1373 struct list_head pbufl; /* partial buffer list */ 1430 * @pfpdu: partial management per user qp 1446 bool partial local [all...] |
/linux-master/crypto/ |
H A D | vmac.c | 62 u8 partial[VMAC_NHBYTES]; /* partial block */ member in union:vmac_desc_ctx::__anon139 65 unsigned int partial_size; /* size of the partial block */ 509 memcpy(&dctx->partial[dctx->partial_size], p, n); 528 memcpy(dctx->partial, p, len); 538 unsigned int partial = dctx->partial_size; local 543 if (partial) { 545 unsigned int n = round_up(partial, 16); 548 memset(&dctx->partial[partial], [all...] |