/linux-master/lib/crypto/ |
H A D | blake2s-generic.c | 39 void blake2s_compress(struct blake2s_state *state, const u8 *block, 43 void blake2s_compress_generic(struct blake2s_state *state, const u8 *block, argument 55 memcpy(m, block, BLAKE2S_BLOCK_SIZE); 105 block += BLAKE2S_BLOCK_SIZE;
|
/linux-master/drivers/media/platform/amphion/ |
H A D | vpu_mbox.c | 37 if (mbox->block) { 59 core->tx_type.block = true; 62 core->tx_data.block = false; 65 core->rx.block = true;
|
/linux-master/fs/gfs2/ |
H A D | rgrp.h | 13 /* Since each block in the file system is represented by two bits in the 66 u64 block); 85 static inline int rgrp_contains_block(struct gfs2_rgrpd *rgd, u64 block) argument 89 return first <= block && block < last;
|
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | amd.c | 390 "for bank %d, block %d (MSR%08X=0x%x%08x)\n", b->cpu, 391 b->bank, b->block, b->address, hi, lo); 405 "for bank %d, block %d (MSR%08X=0x%x%08x)\n", 406 b->cpu, apic, b->bank, b->block, b->address, hi, lo); 519 static u32 smca_get_block_address(unsigned int bank, unsigned int block, argument 522 if (!block) 528 return MSR_AMD64_SMCA_MCx_MISCy(bank, block - 1); 532 unsigned int bank, unsigned int block, 537 if ((bank >= per_cpu(mce_num_banks, cpu)) || (block >= NR_BLOCKS)) 541 return smca_get_block_address(bank, block, cp 531 get_block_address(u32 current_addr, u32 low, u32 high, unsigned int bank, unsigned int block, unsigned int cpu) argument 560 prepare_threshold_block(unsigned int bank, unsigned int block, u32 addr, int offset, u32 misc_high) argument 674 unsigned int bank, block, cpu = smp_processor_id(); local 891 log_and_reset_block(struct threshold_block *block) argument 920 struct threshold_block *first_block = NULL, *block = NULL, *tmp = NULL; local 1120 allocate_threshold_blocks(unsigned int cpu, struct threshold_bank *tb, unsigned int bank, unsigned int block, u32 address) argument [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_audio.h | 32 #define RREG32_ENDPOINT(block, reg) \ 33 radeon_audio_endpoint_rreg(rdev, (block), (reg)) 34 #define WREG32_ENDPOINT(block, reg, v) \ 35 radeon_audio_endpoint_wreg(rdev, (block), (reg), (v))
|
/linux-master/drivers/mtd/nand/raw/ |
H A D | nand_bbt.c | 4 * Bad block table support for the NAND driver 10 * When nand_scan_bbt is called, then it tries to find the bad block table 14 * Once a new bad block is discovered then the "factory" information is updated 21 * If the tables are not versioned, then we "or" the bad block information. 24 * good / bad blocks and the bad block tables are created. 29 * The auto generated bad block table is located in the last good blocks 38 * The table uses 2 bits per block 39 * 11b: block is good 40 * 00b: block is factory marked bad 41 * 01b, 10b: block i 76 bbt_get_entry(struct nand_chip *chip, int block) argument 83 bbt_mark_entry(struct nand_chip *chip, int block, uint8_t mark) argument 558 int startblock, block, dir; local 685 int block = startblock + dir * i; local 716 mark_bbt_block_bad(struct nand_chip *this, struct nand_bbt_descr *td, int chip, int block) argument 784 int block; local 1133 int i, j, chips, block, nrblocks, update; local 1435 int block; local 1449 int block, res; local 1478 int block, ret = 0; local [all...] |
/linux-master/drivers/md/ |
H A D | dm-bufio.c | 315 * Describes how the block was allocated: 331 sector_t block; member in struct:dm_buffer 406 static inline unsigned int cache_index(sector_t block, unsigned int num_locks) argument 408 return dm_hash_locks_index(block, num_locks); 411 static inline void cache_read_lock(struct dm_buffer_cache *bc, sector_t block) argument 414 read_lock_bh(&bc->trees[cache_index(block, bc->num_locks)].u.spinlock); 416 down_read(&bc->trees[cache_index(block, bc->num_locks)].u.lock); 419 static inline void cache_read_unlock(struct dm_buffer_cache *bc, sector_t block) argument 422 read_unlock_bh(&bc->trees[cache_index(block, bc->num_locks)].u.spinlock); 424 up_read(&bc->trees[cache_index(block, b 427 cache_write_lock(struct dm_buffer_cache *bc, sector_t block) argument 435 cache_write_unlock(struct dm_buffer_cache *bc, sector_t block) argument 592 __cache_get(const struct rb_root *root, sector_t block) argument 615 cache_get(struct dm_buffer_cache *bc, sector_t block) argument 899 __find_next(struct rb_root *root, sector_t block) argument 1361 block_to_sector(struct dm_bufio_client *c, sector_t block) argument 1761 __bufio_new(struct dm_bufio_client *c, sector_t block, enum new_flag nf, int *need_submit, struct list_head *write_list) argument 1857 new_read(struct dm_bufio_client *c, sector_t block, enum new_flag nf, struct dm_buffer **bp, unsigned short ioprio) argument 1930 dm_bufio_get(struct dm_bufio_client *c, sector_t block, struct dm_buffer **bp) argument 1937 __dm_bufio_read(struct dm_bufio_client *c, sector_t block, struct dm_buffer **bp, unsigned short ioprio) argument 1946 dm_bufio_read(struct dm_bufio_client *c, sector_t block, struct dm_buffer **bp) argument 1953 dm_bufio_read_with_ioprio(struct dm_bufio_client *c, sector_t block, struct dm_buffer **bp, unsigned short ioprio) argument 1960 dm_bufio_new(struct dm_bufio_client *c, sector_t block, struct dm_buffer **bp) argument 1970 __dm_bufio_prefetch(struct dm_bufio_client *c, sector_t block, unsigned int n_blocks, unsigned short ioprio) argument 2024 dm_bufio_prefetch(struct dm_bufio_client *c, sector_t block, unsigned int n_blocks) argument 2030 dm_bufio_prefetch_with_ioprio(struct dm_bufio_client *c, sector_t block, unsigned int n_blocks, unsigned short ioprio) argument 2207 dm_bufio_issue_discard(struct dm_bufio_client *c, sector_t block, sector_t count) argument 2228 forget_buffer(struct dm_bufio_client *c, sector_t block) argument 2253 dm_bufio_forget(struct dm_bufio_client *c, sector_t block) argument 2266 dm_bufio_forget_buffers(struct dm_bufio_client *c, sector_t block, sector_t n_blocks) argument [all...] |
H A D | dm-dust.c | 78 static int dust_remove_block(struct dust_device *dd, unsigned long long block) argument 84 bblock = dust_rb_search(&dd->badblocklist, block); 88 DMERR("%s: block %llu not found in badblocklist", 89 __func__, block); 98 DMINFO("%s: badblock removed at block %llu", __func__, block); 105 static int dust_add_block(struct dust_device *dd, unsigned long long block, argument 119 bblock->bb = block; 123 DMERR("%s: block %llu already in badblocklist", 124 __func__, block); 141 dust_query_block(struct dust_device *dd, unsigned long long block, char *result, unsigned int maxlen, unsigned int *sz_ptr) argument 421 unsigned long long tmp, block; local [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_init.h | 42 /* Returns the index of start or end of a specific block stage in ops array*/ 43 #define BLOCK_OPS_IDX(block, stage, end) \ 44 (2*(((block)*NUM_OF_INIT_PHASES) + (stage)) + (end)) 490 /* Returns the index of start or end of a specific block stage in ops array */ 491 #define BLOCK_OPS_IDX(block, stage, end) \ 492 (2*(((block)*NUM_OF_INIT_PHASES) + (stage)) + (end)) 539 #define BLOCK_PRTY_INFO(block, en_mask, m1, m1h, m2, m3) \ 541 block##_REG_##block##_PRTY_MASK, \ 542 block##_REG [all...] |
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_btree.h | 34 xfs_bmdr_key_t bmbr; /* bmbt root block */ 44 xfs_bmdr_rec_t bmbr; /* bmbt root block */ 150 /* block allocation / freeing */ 159 const struct xfs_btree_block *block, 163 /* records in block/level */ 338 * Convert from buffer to btree block header. 343 struct xfs_btree_block *block, int level, struct xfs_buf *bp); 348 * Check that block header is ok. 353 struct xfs_btree_block *block, /* generic btree block pointe 427 xfs_btree_get_numrecs(const struct xfs_btree_block *block) argument 432 xfs_btree_set_numrecs(struct xfs_btree_block *block, uint16_t numrecs) argument 438 xfs_btree_get_level(const struct xfs_btree_block *block) argument 633 struct xfs_btree_block *block; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_flower.c | 59 struct mlxsw_sp_flow_block *block, 100 if (mlxsw_sp_flow_block_is_mixed_bound(block)) { 101 NL_SET_ERR_MSG_MOD(extack, "Drop action is not supported when block is bound to ingress and egress"); 104 ingress = mlxsw_sp_flow_block_is_ingress_bound(block); 112 /* Forbid block with this rulei to be bound 134 ruleset = mlxsw_sp_acl_ruleset_lookup(mlxsw_sp, block, 153 if (mlxsw_sp_flow_block_is_egress_bound(block)) { 158 /* Forbid block with this rulei to be bound 196 block, out_dev, 269 block, 58 mlxsw_sp_flower_parse_actions(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct mlxsw_sp_acl_rule_info *rulei, struct flow_action *flow_action, struct netlink_ext_ack *extack) argument 295 mlxsw_sp_flower_parse_meta_iif(struct mlxsw_sp_acl_rule_info *rulei, const struct mlxsw_sp_flow_block *block, const struct flow_match_meta *match, struct netlink_ext_ack *extack) argument 337 mlxsw_sp_flower_parse_meta(struct mlxsw_sp_acl_rule_info *rulei, struct flow_cls_offload *f, struct mlxsw_sp_flow_block *block) argument 557 mlxsw_sp_flower_parse(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct mlxsw_sp_acl_rule_info *rulei, struct flow_cls_offload *f) argument 698 mlxsw_sp_flower_mall_prio_check(struct mlxsw_sp_flow_block *block, struct flow_cls_offload *f) argument 726 mlxsw_sp_flower_replace(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct flow_cls_offload *f) argument 777 mlxsw_sp_flower_destroy(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct flow_cls_offload *f) argument 799 mlxsw_sp_flower_stats(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct flow_cls_offload *f) argument 838 mlxsw_sp_flower_tmplt_create(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct flow_cls_offload *f) argument 859 mlxsw_sp_flower_tmplt_destroy(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct flow_cls_offload *f) argument 875 mlxsw_sp_flower_prio_get(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, u32 chain_index, unsigned int *p_min_prio, unsigned int *p_max_prio) argument [all...] |
/linux-master/fs/omfs/ |
H A D | dir.c | 21 * Finds the bucket for a given name and reads the containing block; 34 static struct buffer_head *omfs_scan_list(struct inode *dir, u64 block, argument 43 while (block != ~0) { 44 bh = omfs_bread(dir->i_sb, block); 51 if (omfs_is_bad(OMFS_SB(dir->i_sb), &oi->i_head, block)) { 59 *prev_block = block; 60 block = be64_to_cpu(oi->i_sibling); 72 u64 block, dummy; local 78 block = be64_to_cpu(*((__be64 *) &bh->b_data[ofs])); 81 return omfs_scan_list(dir, block, nam 118 u64 block; local 165 u64 block, prev; local [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_ttm_buddy_manager.c | 104 struct drm_buddy_block *block; local 106 list_for_each_entry(block, &bman_res->blocks, link) { 108 drm_buddy_block_offset(block) >> PAGE_SHIFT; 112 (drm_buddy_block_size(mm, block) >> PAGE_SHIFT); 160 struct drm_buddy_block *block; local 176 /* Check each drm buddy block individually */ 177 list_for_each_entry(block, &bman_res->blocks, link) { 179 drm_buddy_block_offset(block) >> PAGE_SHIFT; 181 (drm_buddy_block_size(mm, block) >> PAGE_SHIFT); 198 struct drm_buddy_block *block; local 227 struct drm_buddy_block *block; local [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn32/ |
H A D | irq_service_dcn32.c | 201 #define SRI(reg_name, block, id)\ 202 BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 203 reg ## block ## id ## _ ## reg_name 209 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 210 .enable_reg = SRI(reg1, block, reg_num),\ 212 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 214 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 215 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 217 .ack_reg = SRI(reg2, block, reg_num),\ 219 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn35/ |
H A D | irq_service_dcn35.c | 199 #define SRI(reg_name, block, id)\ 200 BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 201 reg ## block ## id ## _ ## reg_name 207 #define IRQ_REG_ENTRY(base, block, reg_num, reg1, mask1, reg2, mask2)\ 208 REG_STRUCT[base + reg_num].enable_reg = SRI(reg1, block, reg_num),\ 210 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 212 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 214 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK, \ 215 REG_STRUCT[base + reg_num].ack_reg = SRI(reg2, block, reg_num),\ 217 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn315/ |
H A D | irq_service_dcn315.c | 207 #define SRI(reg_name, block, id)\ 208 BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 209 reg ## block ## id ## _ ## reg_name 215 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 216 .enable_reg = SRI(reg1, block, reg_num),\ 218 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 220 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 221 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 223 .ack_reg = SRI(reg2, block, reg_num),\ 225 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn302/ |
H A D | irq_service_dcn302.c | 187 #define SRI(reg_name, block, id)\ 188 BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 189 mm ## block ## id ## _ ## reg_name 195 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 196 .enable_reg = SRI(reg1, block, reg_num),\ 197 .enable_mask = block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 199 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 200 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 202 .ack_reg = SRI(reg2, block, reg_num),\ 203 .ack_mask = block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn30/ |
H A D | irq_service_dcn30.c | 212 #define SRI(reg_name, block, id)\ 213 BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 214 mm ## block ## id ## _ ## reg_name 220 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 221 .enable_reg = SRI(reg1, block, reg_num),\ 223 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 225 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 226 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 228 .ack_reg = SRI(reg2, block, reg_num),\ 230 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn31/ |
H A D | irq_service_dcn31.c | 200 #define SRI(reg_name, block, id)\ 201 BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 202 reg ## block ## id ## _ ## reg_name 208 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 209 .enable_reg = SRI(reg1, block, reg_num),\ 211 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 213 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 214 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 216 .ack_reg = SRI(reg2, block, reg_num),\ 218 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn314/ |
H A D | irq_service_dcn314.c | 202 #define SRI(reg_name, block, id)\ 203 (BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 204 reg ## block ## id ## _ ## reg_name) 210 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 211 .enable_reg = SRI(reg1, block, reg_num),\ 213 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 215 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 216 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 218 .ack_reg = SRI(reg2, block, reg_num),\ 220 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn21/ |
H A D | irq_service_dcn21.c | 205 #define SRI(reg_name, block, id)\ 206 BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 207 mm ## block ## id ## _ ## reg_name 213 #define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\ 214 .enable_reg = SRI(reg1, block, reg_num),\ 216 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 218 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 219 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \ 221 .ack_reg = SRI(reg2, block, reg_num),\ 223 block ## reg_nu [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/irq/dcn351/ |
H A D | irq_service_dcn351.c | 178 #define SRI(reg_name, block, id)\ 179 BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 180 reg ## block ## id ## _ ## reg_name 186 #define IRQ_REG_ENTRY(base, block, reg_num, reg1, mask1, reg2, mask2)\ 187 REG_STRUCT[base + reg_num].enable_reg = SRI(reg1, block, reg_num),\ 189 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 191 block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\ 193 ~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK, \ 194 REG_STRUCT[base + reg_num].ack_reg = SRI(reg2, block, reg_num),\ 196 block ## reg_nu [all...] |
/linux-master/drivers/net/ethernet/marvell/prestera/ |
H A D | prestera_acl.c | 24 struct prestera_flow_block *block; member in struct:prestera_acl_ruleset_ht_key 47 struct prestera_counter_block *block; member in struct:prestera_acl_rule_entry::__anon945::__anon949 137 struct prestera_flow_block *block, 144 if (!prestera_acl_chain_is_supported(chain_index, block->ingress)) 152 ruleset->ingress = block->ingress; 153 ruleset->ht_key.block = block; 270 struct prestera_flow_block *block, 276 ht_key.block = block; 136 prestera_acl_ruleset_create(struct prestera_acl *acl, struct prestera_flow_block *block, u32 chain_index) argument 269 __prestera_acl_ruleset_lookup(struct prestera_acl *acl, struct prestera_flow_block *block, u32 chain_index) argument 283 prestera_acl_ruleset_lookup(struct prestera_acl *acl, struct prestera_flow_block *block, u32 chain_index) argument 298 prestera_acl_ruleset_get(struct prestera_acl *acl, struct prestera_flow_block *block, u32 chain_index) argument 345 prestera_acl_ruleset_block_bind(struct prestera_acl_ruleset *ruleset, struct prestera_flow_block *block) argument 369 prestera_acl_ruleset_block_unbind(struct prestera_acl_ruleset *ruleset, struct prestera_flow_block *block) argument 482 struct prestera_flow_block *block = ruleset->ht_key.block; local 533 struct prestera_flow_block *block = ruleset->ht_key.block; local [all...] |
/linux-master/drivers/thunderbolt/ |
H A D | property.c | 38 static struct tb_property_dir *__tb_property_parse_dir(const u32 *block, 95 static struct tb_property *tb_property_parse(const u32 *block, size_t block_len, argument 116 dir = __tb_property_parse_dir(block, block_len, entry->value, 132 parse_dwdata(property->value.data, block + entry->value, 143 parse_dwdata(property->value.text, block + entry->value, 161 static struct tb_property_dir *__tb_property_parse_dir(const u32 *block, argument 177 dir->uuid = kmemdup(&block[dir_offset], sizeof(*dir->uuid), 187 entries = (const struct tb_property_entry *)&block[content_offset]; 195 property = tb_property_parse(block, block_len, &entries[i]); 208 * tb_property_parse_dir() - Parses properties from given property block 220 tb_property_parse_dir(const u32 *block, size_t block_len) argument 348 __tb_property_format_dir(const struct tb_property_dir *dir, u32 *block, unsigned int start_offset, size_t block_len) argument 488 tb_property_format_dir(const struct tb_property_dir *dir, u32 *block, size_t block_len) argument [all...] |
/linux-master/sound/soc/codecs/ |
H A D | tas2781-fmwlib.c | 383 struct tasdev_blk *block) 403 if (block->type == p[i].blktyp) { 413 struct tasdev_blk *block, const struct firmware *fmw, int offset) 426 block->type = be32_to_cpup((__be32 *)&data[offset]); 429 block->is_pchksum_present = data[offset]; 432 block->pchksum = data[offset]; 435 block->is_ychksum_present = data[offset]; 438 block->ychksum = data[offset]; 441 block->blk_size = be32_to_cpup((__be32 *)&data[offset]); 444 block 382 map_dev_idx(struct tasdevice_fw *tas_fmw, struct tasdev_blk *block) argument 412 fw_parse_block_data_kernel(struct tasdevice_fw *tas_fmw, struct tasdev_blk *block, const struct firmware *fmw, int offset) argument 869 tasdevice_load_block_kernel( struct tasdevice_priv *tasdevice, struct tasdev_blk *block) argument 955 fw_parse_block_data(struct tasdevice_fw *tas_fmw, struct tasdev_blk *block, const struct firmware *fmw, int offset) argument 1427 tasdev_bytes_chksum(struct tasdevice_priv *tas_priv, struct tasdev_blk *block, int chn, unsigned char book, unsigned char page, unsigned char reg, unsigned int len, unsigned char val, unsigned char *crc_chksum) argument 1459 tasdev_multibytes_wr(struct tasdevice_priv *tas_priv, struct tasdev_blk *block, int chn, unsigned char book, unsigned char page, unsigned char reg, unsigned char *data, unsigned int len, unsigned int *nr_cmds, unsigned char *crc_chksum) argument 1495 tasdev_block_chksum(struct tasdevice_priv *tas_priv, struct tasdev_blk *block, int chn) argument 1528 tasdev_load_blk(struct tasdevice_priv *tas_priv, struct tasdev_blk *block, int chn) argument 1634 tasdevice_load_block(struct tasdevice_priv *tas_priv, struct tasdev_blk *block) argument 1741 struct tasdev_blk *block; local 2060 struct tasdev_blk *block; local 2140 struct tasdev_blk *block; local [all...] |