Lines Matching refs:chunksize

120  * @chunksize:	Number of bytes of response data expected
136 unsigned int chunksize, u32 stat_pad_len)
159 /* Copy in each dst sg entry from request, up to chunksize */
161 rctx->dst_nents, chunksize);
162 if (datalen < chunksize) {
163 pr_err("%s(): failed to copy dst sg to mbox msg. chunksize %u, datalen %u",
164 __func__, chunksize, datalen);
185 * @chunksize: Number of bytes of request data
199 u8 tx_frag_num, unsigned int chunksize, u32 pad_len)
223 /* Copy in each src sg entry from request, up to chunksize */
225 rctx->src_nents, chunksize);
226 if (unlikely(datalen < chunksize)) {
308 unsigned int chunksize; /* Num bytes of request to submit */
339 chunksize = ctx->max_payload;
341 chunksize = remaining;
343 rctx->src_sent += chunksize;
347 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize);
348 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize);
385 add_to_ctr(rctx->msg_buf.iv_ctr, chunksize >> 4);
395 rctx->src_sent, chunk_start, remaining, chunksize);
403 &cipher_parms, chunksize);
405 atomic64_add(chunksize, &iproc_priv.bytes_out);
407 stat_pad_len = spu->spu_wordalign_padlen(chunksize);
421 dump_sg(rctx->src_sg, rctx->src_skip, chunksize);
439 err = spu_skcipher_rx_sg_create(mssg, rctx, rx_frag_num, chunksize,
453 err = spu_skcipher_tx_sg_create(mssg, rctx, tx_frag_num, chunksize,
604 /* Copy in each src sg entry from request, up to chunksize */
665 unsigned int chunksize = 0; /* length of hash carry + new data */
726 chunksize = nbytes_to_hash;
728 (chunksize > ctx->max_payload))
729 chunksize = ctx->max_payload;
740 rem = chunksize % blocksize; /* remainder */
742 /* chunksize not a multiple of blocksize */
743 chunksize -= rem;
744 if (chunksize == 0) {
765 new_data_len = chunksize - local_nbuf;
782 rctx->total_sent += chunksize;
789 chunksize,
802 atomic64_add(chunksize, &iproc_priv.bytes_out);
812 flow_log("chunk_start: %u chunk_size: %u\n", chunk_start, chunksize);
833 data_pad_len = spu->spu_gcm_ccm_pad_len(ctx->cipher.mode, chunksize);
1101 * Copy in each dst sg entry from request, up to chunksize.
1147 * @chunksize: Number of bytes of request data
1169 unsigned int chunksize,
1211 datalen = chunksize;
1212 if ((chunksize > ctx->digestsize) && incl_icv)
1266 unsigned int chunksize;
1287 chunksize = rctx->total_todo;
1289 flow_log("%s: chunksize %u\n", __func__, chunksize);
1343 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize);
1344 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize);
1351 rctx->total_sent = chunksize;
1352 rctx->src_sent = chunksize;
1371 chunksize);
1389 chunksize - digestsize);
1393 chunksize, rctx->is_encrypt,
1406 aead_parms.assoc_size + chunksize -
1411 aead_parms.assoc_size + chunksize);
1425 atomic64_add(chunksize, &iproc_priv.bytes_out);
1427 flow_log("%s()-sent chunksize:%u\n", __func__, chunksize);
1435 &aead_parms, chunksize);
1439 chunksize, aead_parms.aad_pad_len,
1460 dump_sg(rctx->src_sg, rctx->src_skip, chunksize);
1473 resp_len = chunksize;
1512 assoc_nents, aead_parms.iv_len, chunksize,