/linux-master/arch/x86/crypto/ |
H A D | ecb_cbc_helpers.h | 32 #define ECB_WALK_ADVANCE(blocks) do { \ 33 dst += (blocks) * __bsize; \ 34 src += (blocks) * __bsize; \ 35 nbytes -= (blocks) * __bsize; \ 38 #define ECB_BLOCK(blocks, func) do { \ 39 const int __blocks = (blocks); \ 46 ECB_WALK_ADVANCE(blocks); \ 61 #define CBC_DEC_BLOCK(blocks, func) do { \ 62 const int __blocks = (blocks); \ 68 const u8 *__iv = src + ((blocks) [all...] |
H A D | blake2s-glue.c | 40 const size_t blocks = min_t(size_t, nblocks, local 46 blake2s_compress_avx512(state, block, blocks, inc); 48 blake2s_compress_ssse3(state, block, blocks, inc); 51 nblocks -= blocks; 52 block += blocks * BLAKE2S_BLOCK_SIZE;
|
/linux-master/crypto/async_tx/ |
H A D | async_pq.c | 20 /* the struct page *blocks[] parameter passed to async_gen_syndrome() 22 * blocks[disks-2] and the 'Q' destination address at blocks[disks-1] 107 do_sync_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, argument 117 srcs = (void **) blocks; 120 if (blocks[i] == NULL) { 124 srcs[i] = page_address(blocks[i]) + offsets[i]; 157 * @blocks: source blocks from idx 0..disks-3, P @ disks-2 and Q @ disks-1 159 * @disks: number of blocks (includin 177 async_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, size_t len, struct async_submit_ctl *submit) argument 272 pq_val_chan(struct async_submit_ctl *submit, struct page **blocks, int disks, size_t len) argument 298 async_syndrome_val(struct page **blocks, unsigned int *offsets, int disks, size_t len, enum sum_check_flags *pqres, struct page *spare, unsigned int s_off, struct async_submit_ctl *submit) argument [all...] |
H A D | async_raid6_recov.c | 154 struct page **blocks, unsigned int *offs, 168 p = blocks[disks-2]; 170 q = blocks[disks-1]; 173 a = blocks[faila]; 175 b = blocks[failb]; 204 struct page **blocks, unsigned int *offs, 222 if (blocks[i] == NULL) 231 p = blocks[disks-2]; 233 q = blocks[disks-1]; 235 g = blocks[goo 153 __2data_recov_4(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument 203 __2data_recov_5(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument 294 __2data_recov_n(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument 393 async_raid6_2data_recov(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument 471 async_raid6_datap_recov(int disks, size_t bytes, int faila, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument [all...] |
/linux-master/arch/arm64/crypto/ |
H A D | sha512-ce-glue.c | 30 int blocks); 32 asmlinkage void sha512_block_data_order(u64 *digest, u8 const *src, int blocks); 35 int blocks) 37 while (blocks) { 41 rem = __sha512_ce_transform(sst, src, blocks); 43 src += (blocks - rem) * SHA512_BLOCK_SIZE; 44 blocks = rem; 49 int blocks) 51 sha512_block_data_order(sst->state, src, blocks); 34 sha512_ce_transform(struct sha512_state *sst, u8 const *src, int blocks) argument 48 sha512_arm64_transform(struct sha512_state *sst, u8 const *src, int blocks) argument
|
H A D | sha2-ce-glue.c | 34 int blocks); 37 int blocks) 39 while (blocks) { 45 sst), src, blocks); 47 src += (blocks - rem) * SHA256_BLOCK_SIZE; 48 blocks = rem; 57 asmlinkage void sha256_block_data_order(u32 *digest, u8 const *src, int blocks); 60 int blocks) 62 sha256_block_data_order(sst->state, src, blocks); 36 sha256_ce_transform(struct sha256_state *sst, u8 const *src, int blocks) argument 59 sha256_arm64_transform(struct sha256_state *sst, u8 const *src, int blocks) argument
|
H A D | aes-neonbs-glue.c | 29 int rounds, int blocks); 31 int rounds, int blocks); 34 int rounds, int blocks, u8 iv[]); 37 int rounds, int blocks, u8 iv[]); 40 int rounds, int blocks, u8 iv[]); 42 int rounds, int blocks, u8 iv[]); 46 int rounds, int blocks); 48 int rounds, int blocks, u8 iv[]); 96 int rounds, int blocks)) 106 unsigned int blocks local 94 __ecb_crypt(struct skcipher_request *req, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks)) argument 166 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; local 189 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; local 217 int blocks = (walk.nbytes / AES_BLOCK_SIZE) & ~7; local 276 __xts_crypt(struct skcipher_request *req, bool encrypt, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks, u8 iv[])) argument 316 int blocks = (walk.nbytes / AES_BLOCK_SIZE) & ~7; local [all...] |
H A D | sha3-ce-glue.c | 31 asmlinkage int sha3_ce_transform(u64 *st, const u8 *data, int blocks, 44 int blocks; local 59 blocks = len / sctx->rsiz; 62 while (blocks) { 66 rem = sha3_ce_transform(sctx->st, data, blocks, 69 data += (blocks - rem) * sctx->rsiz; 70 blocks = rem;
|
H A D | sha1-ce-glue.c | 33 int blocks); 36 int blocks) 38 while (blocks) { 44 sst), src, blocks); 46 src += (blocks - rem) * SHA1_BLOCK_SIZE; 47 blocks = rem; 35 sha1_ce_transform(struct sha1_state *sst, u8 const *src, int blocks) argument
|
/linux-master/crypto/ |
H A D | aegis128-core.c | 32 union aegis_block blocks[AEGIS128_STATE_BLOCKS]; member in struct:aegis_state 66 tmp = state->blocks[AEGIS128_STATE_BLOCKS - 1]; 68 crypto_aegis_aesenc(&state->blocks[i], &state->blocks[i - 1], 69 &state->blocks[i]); 70 crypto_aegis_aesenc(&state->blocks[0], &tmp, &state->blocks[0]); 83 crypto_aegis_block_xor(&state->blocks[0], msg); 95 crypto_xor(state->blocks[0].bytes, msg, AEGIS_BLOCK_SIZE); 108 state->blocks[ [all...] |
/linux-master/block/partitions/ |
H A D | sgi.c | 28 __be32 num_blocks; /* Size in logical blocks */ 41 unsigned int start, blocks; local 75 blocks = be32_to_cpu(p->num_blocks); 77 if (blocks) { 78 put_partition(state, slot, start, blocks);
|
/linux-master/fs/xfs/scrub/ |
H A D | rtbitmap_repair.c | 33 unsigned long long blocks = 0; local 36 * Reserve enough blocks to write out a completely new bmbt for a 40 blocks = xfs_bmbt_calc_size(mp, mp->m_sb.sb_rbmblocks); 41 if (blocks > UINT_MAX) 44 rtb->resblks += blocks; 50 * mapped to written blocks. The caller must ensure that the inode is joined 165 unsigned long long blocks = 0; local 177 blocks = xfs_bmbt_calc_size(mp, rtb->rbmblocks); 178 if (blocks > UINT_MAX) 180 if (blocks > rt [all...] |
/linux-master/include/crypto/ |
H A D | sha1_base.h | 19 typedef void (sha1_block_fn)(struct sha1_state *sst, u8 const *src, int blocks); 46 int blocks; local 58 blocks = len / SHA1_BLOCK_SIZE; 61 if (blocks) { 62 block_fn(sctx, data, blocks); 63 data += blocks * SHA1_BLOCK_SIZE;
|
H A D | sm3_base.h | 19 typedef void (sm3_block_fn)(struct sm3_state *sst, u8 const *src, int blocks); 49 int blocks; local 61 blocks = len / SM3_BLOCK_SIZE; 64 if (blocks) { 65 block_fn(sctx, data, blocks); 66 data += blocks * SM3_BLOCK_SIZE;
|
H A D | sha256_base.h | 19 int blocks); 47 int blocks; local 59 blocks = len / SHA256_BLOCK_SIZE; 62 if (blocks) { 63 block_fn(sctx, data, blocks); 64 data += blocks * SHA256_BLOCK_SIZE;
|
H A D | sha512_base.h | 20 int blocks); 69 int blocks; local 81 blocks = len / SHA512_BLOCK_SIZE; 84 if (blocks) { 85 block_fn(sctx, data, blocks); 86 data += blocks * SHA512_BLOCK_SIZE;
|
/linux-master/fs/gfs2/ |
H A D | trans.h | 26 /* reserve either the number of blocks to be allocated plus the rg header 27 * block, or all of the blocks in the rg, whichever is smaller */ 38 unsigned int blocks, unsigned int revokes, 40 int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks,
|
/linux-master/net/mac80211/ |
H A D | michael.c | 57 size_t block, blocks, left; local 63 blocks = data_len / 4; 66 for (block = 0; block < blocks; block++) 75 val |= data[blocks * 4 + left];
|
/linux-master/arch/arm/crypto/ |
H A D | blake2b-neon-glue.c | 30 const size_t blocks = min_t(size_t, nblocks, local 34 blake2b_compress_neon(state, block, blocks, inc); 37 nblocks -= blocks; 38 block += blocks * BLAKE2B_BLOCK_SIZE;
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_ttm_vram_mgr_types.h | 44 /** @blocks: list of DRM buddy blocks */ 45 struct list_head blocks; member in struct:xe_ttm_vram_mgr_resource
|
/linux-master/arch/m68k/emu/ |
H A D | nfblock.c | 40 static inline s32 nfhd_get_capacity(u32 major, u32 minor, u32 *blocks, argument 44 virt_to_phys(blocks), virt_to_phys(blocksize)); 55 u32 blocks, bsize; member in struct:nfhd_device 84 geo->cylinders = dev->blocks >> (6 - dev->bshift); 97 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) argument 106 pr_info("nfhd%u: found device with %u blocks (%u bytes)\n", dev_id, 107 blocks, bsize); 119 dev->blocks = blocks; 135 set_capacity(dev->disk, (sector_t)blocks * (bsiz 154 u32 blocks, bsize; local [all...] |
/linux-master/mm/ |
H A D | dmapool_test.c | 40 static int dmapool_test_alloc(struct dma_pool_pair *p, int blocks) argument 44 for (i = 0; i < blocks; i++) { 51 for (i = 0; i < blocks; i++) 64 int blocks = nr_blocks(parms->size); local 69 p = kcalloc(blocks, sizeof(*p), GFP_KERNEL); 82 ret = dmapool_test_alloc(p, blocks); 90 printk("dmapool test: size:%-4zu align:%-4zu blocks:%-4d time:%llu\n", 91 parms->size, parms->align, blocks,
|
/linux-master/drivers/mtd/ |
H A D | rfd_ftl.c | 88 struct block *blocks; member in struct:partition 95 struct block *block = &part->blocks[block_no]; 188 part->blocks = kcalloc(part->total_blocks, sizeof(struct block), 190 if (!part->blocks) 235 kfree(part->blocks); 277 erase->addr = part->blocks[block].offset; 280 part->blocks[block].state = BLOCK_ERASING; 281 part->blocks[block].free_sectors = 0; 288 part->blocks[block].state = BLOCK_FAILED; 289 part->blocks[bloc [all...] |
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_btree_staging.c | 27 * initializing new btree blocks and filling them with records or key/ptr 183 * height of and the number of blocks needed to construct the btree. See the 187 * In step four, the caller must allocate xfs_btree_bload.nr_blocks blocks and 189 * blocks to be allocated beforehand to avoid ENOSPC failures midway through a 196 * is responsible for cleaning up the previous btree blocks, if any. 204 * is the number of blocks in the next lower level of the tree. For each 209 * The number of blocks for the level is defined to be: 211 * blocks = floor(nr_items / desired) 217 * npb = nr_items / blocks 219 * Some of the leftmost blocks i 480 xfs_btree_bload_level_geometry( struct xfs_btree_cur *cur, struct xfs_btree_bload *bbl, unsigned int level, uint64_t nr_this_level, unsigned int *avg_per_block, uint64_t *blocks, uint64_t *blocks_with_extra) argument 681 uint64_t blocks; local [all...] |
/linux-master/drivers/iio/buffer/ |
H A D | industrialio-buffer-dma.c | 21 * For DMA buffers the storage is sub-divided into so called blocks. Each block 53 * In addition to this blocks are reference counted and the memory associated 59 * managing the blocks. 222 * @queue: Queue for which to complete blocks. 223 * @list: List of aborted blocks. All blocks in this list must be from @queue. 227 * hand the blocks back to the queue. 298 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { 299 block = queue->fileio.blocks[i]; 307 * At this point all blocks ar [all...] |