Searched refs:blocks (Results 1 - 25 of 448) sorted by relevance

1234567891011>>

/linux-master/arch/x86/crypto/
H A Decb_cbc_helpers.h32 #define ECB_WALK_ADVANCE(blocks) do { \
33 dst += (blocks) * __bsize; \
34 src += (blocks) * __bsize; \
35 nbytes -= (blocks) * __bsize; \
38 #define ECB_BLOCK(blocks, func) do { \
39 const int __blocks = (blocks); \
46 ECB_WALK_ADVANCE(blocks); \
61 #define CBC_DEC_BLOCK(blocks, func) do { \
62 const int __blocks = (blocks); \
68 const u8 *__iv = src + ((blocks)
[all...]
H A Dblake2s-glue.c40 const size_t blocks = min_t(size_t, nblocks, local
46 blake2s_compress_avx512(state, block, blocks, inc);
48 blake2s_compress_ssse3(state, block, blocks, inc);
51 nblocks -= blocks;
52 block += blocks * BLAKE2S_BLOCK_SIZE;
/linux-master/crypto/async_tx/
H A Dasync_pq.c20 /* the struct page *blocks[] parameter passed to async_gen_syndrome()
22 * blocks[disks-2] and the 'Q' destination address at blocks[disks-1]
107 do_sync_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, argument
117 srcs = (void **) blocks;
120 if (blocks[i] == NULL) {
124 srcs[i] = page_address(blocks[i]) + offsets[i];
157 * @blocks: source blocks from idx 0..disks-3, P @ disks-2 and Q @ disks-1
159 * @disks: number of blocks (includin
177 async_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, size_t len, struct async_submit_ctl *submit) argument
272 pq_val_chan(struct async_submit_ctl *submit, struct page **blocks, int disks, size_t len) argument
298 async_syndrome_val(struct page **blocks, unsigned int *offsets, int disks, size_t len, enum sum_check_flags *pqres, struct page *spare, unsigned int s_off, struct async_submit_ctl *submit) argument
[all...]
H A Dasync_raid6_recov.c154 struct page **blocks, unsigned int *offs,
168 p = blocks[disks-2];
170 q = blocks[disks-1];
173 a = blocks[faila];
175 b = blocks[failb];
204 struct page **blocks, unsigned int *offs,
222 if (blocks[i] == NULL)
231 p = blocks[disks-2];
233 q = blocks[disks-1];
235 g = blocks[goo
153 __2data_recov_4(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument
203 __2data_recov_5(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument
294 __2data_recov_n(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument
393 async_raid6_2data_recov(int disks, size_t bytes, int faila, int failb, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument
471 async_raid6_datap_recov(int disks, size_t bytes, int faila, struct page **blocks, unsigned int *offs, struct async_submit_ctl *submit) argument
[all...]
/linux-master/arch/arm64/crypto/
H A Dsha512-ce-glue.c30 int blocks);
32 asmlinkage void sha512_block_data_order(u64 *digest, u8 const *src, int blocks);
35 int blocks)
37 while (blocks) {
41 rem = __sha512_ce_transform(sst, src, blocks);
43 src += (blocks - rem) * SHA512_BLOCK_SIZE;
44 blocks = rem;
49 int blocks)
51 sha512_block_data_order(sst->state, src, blocks);
34 sha512_ce_transform(struct sha512_state *sst, u8 const *src, int blocks) argument
48 sha512_arm64_transform(struct sha512_state *sst, u8 const *src, int blocks) argument
H A Dsha2-ce-glue.c34 int blocks);
37 int blocks)
39 while (blocks) {
45 sst), src, blocks);
47 src += (blocks - rem) * SHA256_BLOCK_SIZE;
48 blocks = rem;
57 asmlinkage void sha256_block_data_order(u32 *digest, u8 const *src, int blocks);
60 int blocks)
62 sha256_block_data_order(sst->state, src, blocks);
36 sha256_ce_transform(struct sha256_state *sst, u8 const *src, int blocks) argument
59 sha256_arm64_transform(struct sha256_state *sst, u8 const *src, int blocks) argument
H A Daes-neonbs-glue.c29 int rounds, int blocks);
31 int rounds, int blocks);
34 int rounds, int blocks, u8 iv[]);
37 int rounds, int blocks, u8 iv[]);
40 int rounds, int blocks, u8 iv[]);
42 int rounds, int blocks, u8 iv[]);
46 int rounds, int blocks);
48 int rounds, int blocks, u8 iv[]);
96 int rounds, int blocks))
106 unsigned int blocks local
94 __ecb_crypt(struct skcipher_request *req, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks)) argument
166 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; local
189 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; local
217 int blocks = (walk.nbytes / AES_BLOCK_SIZE) & ~7; local
276 __xts_crypt(struct skcipher_request *req, bool encrypt, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks, u8 iv[])) argument
316 int blocks = (walk.nbytes / AES_BLOCK_SIZE) & ~7; local
[all...]
H A Dsha3-ce-glue.c31 asmlinkage int sha3_ce_transform(u64 *st, const u8 *data, int blocks,
44 int blocks; local
59 blocks = len / sctx->rsiz;
62 while (blocks) {
66 rem = sha3_ce_transform(sctx->st, data, blocks,
69 data += (blocks - rem) * sctx->rsiz;
70 blocks = rem;
H A Dsha1-ce-glue.c33 int blocks);
36 int blocks)
38 while (blocks) {
44 sst), src, blocks);
46 src += (blocks - rem) * SHA1_BLOCK_SIZE;
47 blocks = rem;
35 sha1_ce_transform(struct sha1_state *sst, u8 const *src, int blocks) argument
/linux-master/crypto/
H A Daegis128-core.c32 union aegis_block blocks[AEGIS128_STATE_BLOCKS]; member in struct:aegis_state
66 tmp = state->blocks[AEGIS128_STATE_BLOCKS - 1];
68 crypto_aegis_aesenc(&state->blocks[i], &state->blocks[i - 1],
69 &state->blocks[i]);
70 crypto_aegis_aesenc(&state->blocks[0], &tmp, &state->blocks[0]);
83 crypto_aegis_block_xor(&state->blocks[0], msg);
95 crypto_xor(state->blocks[0].bytes, msg, AEGIS_BLOCK_SIZE);
108 state->blocks[
[all...]
/linux-master/block/partitions/
H A Dsgi.c28 __be32 num_blocks; /* Size in logical blocks */
41 unsigned int start, blocks; local
75 blocks = be32_to_cpu(p->num_blocks);
77 if (blocks) {
78 put_partition(state, slot, start, blocks);
/linux-master/fs/xfs/scrub/
H A Drtbitmap_repair.c33 unsigned long long blocks = 0; local
36 * Reserve enough blocks to write out a completely new bmbt for a
40 blocks = xfs_bmbt_calc_size(mp, mp->m_sb.sb_rbmblocks);
41 if (blocks > UINT_MAX)
44 rtb->resblks += blocks;
50 * mapped to written blocks. The caller must ensure that the inode is joined
165 unsigned long long blocks = 0; local
177 blocks = xfs_bmbt_calc_size(mp, rtb->rbmblocks);
178 if (blocks > UINT_MAX)
180 if (blocks > rt
[all...]
/linux-master/include/crypto/
H A Dsha1_base.h19 typedef void (sha1_block_fn)(struct sha1_state *sst, u8 const *src, int blocks);
46 int blocks; local
58 blocks = len / SHA1_BLOCK_SIZE;
61 if (blocks) {
62 block_fn(sctx, data, blocks);
63 data += blocks * SHA1_BLOCK_SIZE;
H A Dsm3_base.h19 typedef void (sm3_block_fn)(struct sm3_state *sst, u8 const *src, int blocks);
49 int blocks; local
61 blocks = len / SM3_BLOCK_SIZE;
64 if (blocks) {
65 block_fn(sctx, data, blocks);
66 data += blocks * SM3_BLOCK_SIZE;
H A Dsha256_base.h19 int blocks);
47 int blocks; local
59 blocks = len / SHA256_BLOCK_SIZE;
62 if (blocks) {
63 block_fn(sctx, data, blocks);
64 data += blocks * SHA256_BLOCK_SIZE;
H A Dsha512_base.h20 int blocks);
69 int blocks; local
81 blocks = len / SHA512_BLOCK_SIZE;
84 if (blocks) {
85 block_fn(sctx, data, blocks);
86 data += blocks * SHA512_BLOCK_SIZE;
/linux-master/fs/gfs2/
H A Dtrans.h26 /* reserve either the number of blocks to be allocated plus the rg header
27 * block, or all of the blocks in the rg, whichever is smaller */
38 unsigned int blocks, unsigned int revokes,
40 int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks,
/linux-master/net/mac80211/
H A Dmichael.c57 size_t block, blocks, left; local
63 blocks = data_len / 4;
66 for (block = 0; block < blocks; block++)
75 val |= data[blocks * 4 + left];
/linux-master/arch/arm/crypto/
H A Dblake2b-neon-glue.c30 const size_t blocks = min_t(size_t, nblocks, local
34 blake2b_compress_neon(state, block, blocks, inc);
37 nblocks -= blocks;
38 block += blocks * BLAKE2B_BLOCK_SIZE;
/linux-master/drivers/gpu/drm/xe/
H A Dxe_ttm_vram_mgr_types.h44 /** @blocks: list of DRM buddy blocks */
45 struct list_head blocks; member in struct:xe_ttm_vram_mgr_resource
/linux-master/arch/m68k/emu/
H A Dnfblock.c40 static inline s32 nfhd_get_capacity(u32 major, u32 minor, u32 *blocks, argument
44 virt_to_phys(blocks), virt_to_phys(blocksize));
55 u32 blocks, bsize; member in struct:nfhd_device
84 geo->cylinders = dev->blocks >> (6 - dev->bshift);
97 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) argument
106 pr_info("nfhd%u: found device with %u blocks (%u bytes)\n", dev_id,
107 blocks, bsize);
119 dev->blocks = blocks;
135 set_capacity(dev->disk, (sector_t)blocks * (bsiz
154 u32 blocks, bsize; local
[all...]
/linux-master/mm/
H A Ddmapool_test.c40 static int dmapool_test_alloc(struct dma_pool_pair *p, int blocks) argument
44 for (i = 0; i < blocks; i++) {
51 for (i = 0; i < blocks; i++)
64 int blocks = nr_blocks(parms->size); local
69 p = kcalloc(blocks, sizeof(*p), GFP_KERNEL);
82 ret = dmapool_test_alloc(p, blocks);
90 printk("dmapool test: size:%-4zu align:%-4zu blocks:%-4d time:%llu\n",
91 parms->size, parms->align, blocks,
/linux-master/drivers/mtd/
H A Drfd_ftl.c88 struct block *blocks; member in struct:partition
95 struct block *block = &part->blocks[block_no];
188 part->blocks = kcalloc(part->total_blocks, sizeof(struct block),
190 if (!part->blocks)
235 kfree(part->blocks);
277 erase->addr = part->blocks[block].offset;
280 part->blocks[block].state = BLOCK_ERASING;
281 part->blocks[block].free_sectors = 0;
288 part->blocks[block].state = BLOCK_FAILED;
289 part->blocks[bloc
[all...]
/linux-master/fs/xfs/libxfs/
H A Dxfs_btree_staging.c27 * initializing new btree blocks and filling them with records or key/ptr
183 * height of and the number of blocks needed to construct the btree. See the
187 * In step four, the caller must allocate xfs_btree_bload.nr_blocks blocks and
189 * blocks to be allocated beforehand to avoid ENOSPC failures midway through a
196 * is responsible for cleaning up the previous btree blocks, if any.
204 * is the number of blocks in the next lower level of the tree. For each
209 * The number of blocks for the level is defined to be:
211 * blocks = floor(nr_items / desired)
217 * npb = nr_items / blocks
219 * Some of the leftmost blocks i
480 xfs_btree_bload_level_geometry( struct xfs_btree_cur *cur, struct xfs_btree_bload *bbl, unsigned int level, uint64_t nr_this_level, unsigned int *avg_per_block, uint64_t *blocks, uint64_t *blocks_with_extra) argument
681 uint64_t blocks; local
[all...]
/linux-master/drivers/iio/buffer/
H A Dindustrialio-buffer-dma.c21 * For DMA buffers the storage is sub-divided into so called blocks. Each block
53 * In addition to this blocks are reference counted and the memory associated
59 * managing the blocks.
222 * @queue: Queue for which to complete blocks.
223 * @list: List of aborted blocks. All blocks in this list must be from @queue.
227 * hand the blocks back to the queue.
298 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) {
299 block = queue->fileio.blocks[i];
307 * At this point all blocks ar
[all...]

Completed in 368 milliseconds

1234567891011>>