/linux-master/tools/testing/selftests/syscall_user_dispatch/ |
H A D | sud_benchmark.c | 97 double partial = 0; local 100 partial += one_sysinfo_step()/(CALIBRATION_STEP*factor); 101 return partial;
|
/linux-master/include/net/ |
H A D | espintcp.h | 25 struct espintcp_msg partial; member in struct:espintcp_ctx
|
H A D | gso.h | 49 /* Do not update partial checksums if remote checksum is enabled. */ 69 __wsum partial = SKB_GSO_CB(skb)->csum; local 74 return csum_fold(csum_partial(csum_start, plen, partial));
|
/linux-master/include/keys/ |
H A D | asymmetric-type.h | 85 bool partial);
|
/linux-master/arch/powerpc/crypto/ |
H A D | aes-gcm-p10.S | 637 # check partial block 888 std 15, 56(7) # clear partial? 927 # check partial block 929 ld 15, 56(7) # partial? 943 # create partial block mask 955 lxvb16x 16, 15, 10 # load partial block mask 965 std 12, 56(7) # update partial? 986 lxvb16x \_mask, 0, 10 # load partial block mask 990 # Handle multiple partial blocks for encrypt and decrypt 1029 # store partial bloc [all...] |
/linux-master/arch/s390/lib/ |
H A D | Makefile | 7 lib-y += csum-partial.o
|
/linux-master/arch/x86/kernel/ |
H A D | dumpstack.c | 155 bool partial, const char *log_lvl) 166 if (!partial && on_stack(info, regs, sizeof(*regs))) { 169 } else if (partial && on_stack(info, (void *)regs + IRET_FRAME_OFFSET, 193 bool partial = false; local 198 regs = unwind_get_entry_regs(&state, &partial); 240 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); 299 regs = unwind_get_entry_regs(&state, &partial); 301 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); 154 show_regs_if_on_stack(struct stack_info *info, struct pt_regs *regs, bool partial, const char *log_lvl) argument
|
/linux-master/crypto/ |
H A D | sm3.c | 180 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; local 185 if ((partial + len) >= SM3_BLOCK_SIZE) { 188 if (partial) { 189 int p = SM3_BLOCK_SIZE - partial; 191 memcpy(sctx->buffer + partial, data, p); 208 partial = 0; 211 memcpy(sctx->buffer + partial, data, len); 220 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; local 224 sctx->buffer[partial++] = 0x80; 225 if (partial > bit_offse [all...] |
H A D | vmac.c | 62 u8 partial[VMAC_NHBYTES]; /* partial block */ member in union:vmac_desc_ctx::__anon139 65 unsigned int partial_size; /* size of the partial block */ 509 memcpy(&dctx->partial[dctx->partial_size], p, n); 528 memcpy(dctx->partial, p, len); 538 unsigned int partial = dctx->partial_size; local 543 if (partial) { 545 unsigned int n = round_up(partial, 16); 548 memset(&dctx->partial[partial], [all...] |
/linux-master/drivers/crypto/stm32/ |
H A D | stm32-crc32.c | 69 u32 partial; /* crc32c: partial in first 4 bytes of that struct */ member in struct:stm32_crc_desc_ctx 136 /* Store partial result */ 137 ctx->partial = readl_relaxed(crc->regs + CRC_DR); 163 ctx->partial = crc32_le(ctx->partial, d8, length); 165 ctx->partial = __crc32c_le(ctx->partial, d8, length); 176 writel_relaxed(bitrev32(ctx->partial), crc->regs + CRC_INIT); 205 /* Store partial resul [all...] |
/linux-master/arch/arm64/crypto/ |
H A D | ghash-ce-glue.c | 122 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; local 126 if ((partial + len) >= GHASH_BLOCK_SIZE) { 130 if (partial) { 131 int p = GHASH_BLOCK_SIZE - partial; 133 memcpy(ctx->buf + partial, src, p); 145 partial ? ctx->buf : NULL, 150 partial = 0; 154 memcpy(ctx->buf + partial, src, len); 161 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; local 163 if (partial) { [all...] |
/linux-master/fs/ocfs2/ |
H A D | aops.h | 26 int *partial,
|
/linux-master/lib/crypto/ |
H A D | chacha20poly1305.c | 219 size_t partial = 0; local 265 if (unlikely(partial)) { 266 size_t l = min(length, CHACHA_BLOCK_SIZE - partial); 268 crypto_xor(addr, b.chacha_stream + partial, l); 269 partial = (partial + l) & (CHACHA_BLOCK_SIZE - 1); 289 partial = length;
|
/linux-master/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
H A D | ppevvmath.h | 38 * A.partial.real => Only the integer portion 39 * A.partial.decimal => Only the fractional portion 46 } partial; member in union:_fInt 341 X_LessThanOne = (X.partial.real == 0 && X.partial.decimal != 0 && X.full >= 0); 342 Y_LessThanOne = (Y.partial.real == 0 && Y.partial.decimal != 0 && Y.full >= 0); 413 if (num.partial.real > 3000) 415 else if (num.partial.real > 1000) 417 else if (num.partial [all...] |
/linux-master/drivers/iommu/ |
H A D | io-pgfault.c | 99 list_add(&iopf->list, &fault_param->partial); 127 /* See if we have partial faults for this group */ 129 list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { 203 * leaving, otherwise partial faults will be stuck. 291 * iopf_queue_discard_partial - Remove all pending partial fault 292 * @queue: the queue whose partial faults need to be discarded 295 * lost and the IOMMU driver calls this to discard all partial faults. The 311 list_for_each_entry_safe(iopf, next, &iopf_param->partial, 356 INIT_LIST_HEAD(&fault_param->partial); 413 list_for_each_entry_safe(partial_iopf, next, &fault_param->partial, lis [all...] |
/linux-master/tools/mm/ |
H A D | slabinfo.c | 36 unsigned long partial, objects, slabs, objects_partial, objects_total; member in struct:slabinfo 128 "-P|--partial Sort by number of partial slabs\n" 500 printf("Add partial %8lu %8lu %3lu %3lu\n", 505 printf("Remove partial %8lu %8lu %3lu %3lu\n", 510 printf("Cpu partial list %8lu %8lu %3lu %3lu\n", 535 printf("Moved to head of partial list %7lu %3lu%%\n", 537 printf("Moved to tail of partial list %7lu %3lu%%\n", 576 s->slab_size, s->slabs - s->partial - s->cpu_slabs, 579 page_size << s->order, s->partial, onof 1352 _xtotals(char *heading, char *underline, int loss, int size, int partial) argument [all...] |
/linux-master/net/netfilter/ |
H A D | xt_dccp.c | 56 goto partial; 74 partial:
|
/linux-master/include/linux/ |
H A D | shdma-base.h | 50 size_t partial; member in struct:shdma_desc
|
H A D | splice.h | 59 struct partial_page *partial; /* pages[] may not be contig */ member in struct:splice_pipe_desc 61 unsigned int nr_pages_max; /* pages[] & partial[] arrays size */
|
/linux-master/arch/arm/crypto/ |
H A D | ghash-ce-glue.c | 115 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; local 119 if ((partial + len) >= GHASH_BLOCK_SIZE) { 123 if (partial) { 124 int p = GHASH_BLOCK_SIZE - partial; 126 memcpy(ctx->buf + partial, src, p); 135 partial ? ctx->buf : NULL); 137 partial = 0; 140 memcpy(ctx->buf + partial, src, len); 147 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; local 149 if (partial) { [all...] |
/linux-master/fs/reiserfs/ |
H A D | file.c | 174 int partial = 0; local 204 partial = 1; 229 * If this is a partial write which happened to make all buffers 232 * uptodate as a result of this (potentially partial) write. 234 if (!partial)
|
/linux-master/drivers/usb/storage/ |
H A D | transport.c | 257 unsigned int length, int result, unsigned int partial) 260 result, partial, length); 265 if (partial != length) { 469 unsigned int partial; local 472 &partial); 474 scsi_set_resid(srb, scsi_bufflen(srb) - partial); 492 unsigned int partial; local 499 length_left, &partial); 500 length_left -= partial; 504 length_left, &partial); 256 interpret_urb_result(struct us_data *us, unsigned int pipe, unsigned int length, int result, unsigned int partial) argument [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma.h | 245 offsetof(typeof(*view), partial)); 248 return memcmp(&vma->gtt_view.partial, &view->partial, view->type);
|
H A D | i915_vma_types.h | 167 struct intel_partial_info partial; member in union:i915_gtt_view::__anon446
|
/linux-master/fs/ext4/ |
H A D | extents.c | 2407 * file system when freeing a partial cluster containing at least one 2408 * delayed and unwritten block. A partial cluster meeting that 2432 struct partial_cluster *partial, 2463 trace_ext4_remove_blocks(inode, ex, from, to, partial); 2466 * if we have a partial cluster, and it's different from the 2471 if (partial->state != initial && 2472 partial->pclu != EXT4_B2C(sbi, last_pblk)) { 2473 if (partial->state == tofree) { 2475 if (ext4_is_pending(inode, partial->lblk)) 2478 EXT4_C2B(sbi, partial 2430 ext4_remove_blocks(handle_t *handle, struct inode *inode, struct ext4_extent *ex, struct partial_cluster *partial, ext4_lblk_t from, ext4_lblk_t to) argument 2566 ext4_ext_rm_leaf(handle_t *handle, struct inode *inode, struct ext4_ext_path *path, struct partial_cluster *partial, ext4_lblk_t start, ext4_lblk_t end) argument 2795 struct partial_cluster partial; local [all...] |