/linux-master/drivers/crypto/ccree/ |
H A D | cc_buffer_mgr.c | 117 u32 buff_size, u32 *curr_nents, 124 new_nents = (*curr_nents + buff_size / CC_MAX_MLLI_ENTRY_SIZE + 1); 132 while (buff_size > CC_MAX_MLLI_ENTRY_SIZE) { 139 buff_size -= CC_MAX_MLLI_ENTRY_SIZE; 145 cc_lli_set_size(mlli_entry_p, buff_size); 116 cc_render_buff_to_mlli(struct device *dev, dma_addr_t buff_dma, u32 buff_size, u32 *curr_nents, u32 **mlli_entry_pp) argument
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_admin.c | 454 size_t buff_size) 469 req.init_cfg_sz = buff_size; 453 adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size) argument
|
H A D | adf_admin.h | 24 int adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size);
|
H A D | adf_gen4_pm_debugfs.c | 113 u32 *pm_info_regs, size_t buff_size, int table_len, 126 wr += scnprintf(&buff[wr], buff_size - wr, "%s: %#x\n", key, 135 u32 *pm_info_regs, size_t buff_size, 138 return pm_scnprint_table(buff, table, pm_info_regs, buff_size, 143 u32 *pm_info_regs, size_t buff_size, 146 return pm_scnprint_table(buff, table, pm_info_regs, buff_size, 112 pm_scnprint_table(char *buff, struct pm_status_row *table, u32 *pm_info_regs, size_t buff_size, int table_len, bool lowercase) argument 134 pm_scnprint_table_upper_keys(char *buff, struct pm_status_row *table, u32 *pm_info_regs, size_t buff_size, int table_len) argument 142 pm_scnprint_table_lower_keys(char *buff, struct pm_status_row *table, u32 *pm_info_regs, size_t buff_size, int table_len) argument
|
/linux-master/drivers/dma/ |
H A D | mv_xor_v2.c | 100 * @buff_size: amount of bytes to be processed 129 u32 buff_size; member in struct:mv_xor_v2_descriptor 421 hw_descriptor->buff_size = len; 480 hw_descriptor->buff_size = len;
|
/linux-master/drivers/firmware/efi/libstub/ |
H A D | efi-stub-helper.c | 462 map->map_size = map->buff_size;
|
H A D | mem.c | 54 m->buff_size = m->map_size = size;
|
/linux-master/drivers/gpu/drm/amd/display/dc/dsc/ |
H A D | dc_dsc.c | 337 int buff_size; local 344 buff_size = dpcd_dsc_basic_data[DP_DSC_RC_BUF_SIZE - DP_DSC_SUPPORT] + 1; 345 dsc_sink_caps->rc_buffer_size = buff_size * buff_block_size;
|
/linux-master/drivers/gpu/drm/bridge/cadence/ |
H A D | cdns-mhdp8546-core.c | 144 u8 *buff, u16 buff_size) 149 for (i = 0; i < buff_size; i++) { 143 cdns_mhdp_mailbox_recv_data(struct cdns_mhdp_device *mhdp, u8 *buff, u16 buff_size) argument
|
H A D | cdns-mhdp8546-hdcp.c | 83 u8 *buff, u16 buff_size) 88 for (i = 0; i < buff_size; i++) { 82 cdns_mhdp_secure_mailbox_recv_data(struct cdns_mhdp_device *mhdp, u8 *buff, u16 buff_size) argument
|
/linux-master/drivers/gpu/drm/rockchip/ |
H A D | cdn-dp-reg.c | 142 u8 *buff, u16 buff_size) 147 for (i = 0; i < buff_size; i++) { 141 cdn_dp_mailbox_read_receive(struct cdn_dp_device *dp, u8 *buff, u16 buff_size) argument
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_device.h | 618 u32 buff_size; member in struct:hns_roce_qp
|
H A D | hns_roce_qp.c | 652 hr_qp->buff_size = 0; 662 hr_qp->buff_size += buf_size; 666 hr_qp->sge.offset = hr_qp->buff_size; 673 hr_qp->buff_size += buf_size; 677 hr_qp->rq.offset = hr_qp->buff_size; 684 hr_qp->buff_size += buf_size; 687 if (hr_qp->buff_size < 1)
|
/linux-master/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_nic.c | 615 unsigned int buff_size = 0U; local 620 buff_size = AQ_CFG_TX_FRAME_MAX; 622 buff_size = frag_len; 625 buff_size, DMA_TO_DEVICE); 634 dx_buff->len = buff_size; 639 frag_len -= buff_size; 640 buff_offset += buff_size; 759 unsigned int buff_size = 0U; local 767 buff_size = AQ_CFG_TX_FRAME_MAX; 769 buff_size [all...] |
/linux-master/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
H A D | hw_atl_a0.c | 128 u32 buff_size = 0U; local 148 buff_size = HW_ATL_A0_TXBUF_MAX; 150 hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, buff_size, tc); 152 (buff_size * 156 (buff_size * 163 buff_size = HW_ATL_A0_RXBUF_MAX; 165 hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, buff_size, tc); 167 (buff_size * 171 (buff_size *
|
/linux-master/drivers/net/ethernet/ibm/ |
H A D | ibmveth.c | 154 u32 buff_size, u32 pool_active) 158 pool->buff_size = buff_size; 229 skb = netdev_alloc_skb(adapter->netdev, pool->buff_size); 248 pool->buff_size, DMA_FROM_DEVICE); 260 desc.fields.flags_len = IBMVETH_BUF_VALID | pool->buff_size; 264 unsigned int len = min(pool->buff_size, 293 pool->dma_addr[index], pool->buff_size, 347 pool->buff_size, 386 adapter->rx_buff_pool[pool].buff_size, 152 ibmveth_init_buffer_pool(struct ibmveth_buff_pool *pool, u32 pool_index, u32 pool_size, u32 buff_size, u32 pool_active) argument [all...] |
H A D | ibmveth.h | 115 u32 buff_size; member in struct:ibmveth_buff_pool
|
H A D | ibmvnic.c | 571 * @buff_size: Size of each buffer in the LTB 573 * Allocate a set of LTBs to accommodate @num_buffs buffers of @buff_size 584 int buff_size) 597 dev_dbg(dev, "%s() num_buffs %d, buff_size %d\n", __func__, num_buffs, 598 buff_size); 600 ltb_size = rounddown(IBMVNIC_ONE_LTB_SIZE, buff_size); 601 tot_size = num_buffs * buff_size; 703 nbufs = ltb->size / rxpool->buff_size; 710 *offset = bufidx * rxpool->buff_size; 797 pool->buff_size); 582 alloc_ltb_set(struct ibmvnic_adapter *adapter, struct ibmvnic_ltb_set *ltb_set, int num_buffs, int buff_size) argument 1060 u64 buff_size; local 1300 u64 buff_size; local [all...] |
H A D | ibmvnic.h | 872 int buff_size; member in struct:ibmvnic_rx_pool
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_adminq.c | 744 * @buff_size: size of buffer for indirect commands 755 u16 buff_size, 806 if (buff_size > hw->aq.asq_buf_size) { 810 buff_size); 848 memcpy(dma_buff->va, buff, buff_size); 849 desc_on_ring->datalen = cpu_to_le16(buff_size); 863 buff, buff_size); 896 memcpy(buff, dma_buff->va, buff_size); 919 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); 948 * @buff_size 752 i40e_asq_send_command_atomic_exec(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, bool is_atomic_context) argument 956 i40e_asq_send_command_atomic(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, bool is_atomic_context) argument 975 i40e_asq_send_command(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details) argument 998 i40e_asq_send_command_atomic_v2(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, bool is_atomic_context, enum i40e_admin_queue_err *aq_status) argument 1020 i40e_asq_send_command_v2(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, enum i40e_admin_queue_err *aq_status) argument [all...] |
H A D | i40e_common.c | 3081 * @buff_size: Size of the virtual buffer 3089 void *buff, u16 buff_size, u16 *data_size, 3108 if (buff_size > I40E_AQ_LARGE_BUF) 3111 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); 3223 * @buff_size: size of the buffer (in bytes) 3231 u8 mib_type, void *buff, u16 buff_size, 3242 if (buff_size == 0 || !buff) 3252 desc.datalen = cpu_to_le16(buff_size); 3255 if (buff_size > I40E_AQ_LARGE_BUF) 3258 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_detail 3088 i40e_aq_discover_capabilities(struct i40e_hw *hw, void *buff, u16 buff_size, u16 *data_size, enum i40e_admin_queue_opc list_type_opc, struct i40e_asq_cmd_details *cmd_details) argument 3230 i40e_aq_get_lldp_mib(struct i40e_hw *hw, u8 bridge_type, u8 mib_type, void *buff, u16 buff_size, u16 *local_len, u16 *remote_len, struct i40e_asq_cmd_details *cmd_details) argument 3280 i40e_aq_set_lldp_mib(struct i40e_hw *hw, u8 mib_type, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details) argument 3484 i40e_aq_get_cee_dcb_config(struct i40e_hw *hw, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details) argument 3625 i40e_aq_tx_sched_cmd(struct i40e_hw *hw, u16 seid, void *buff, u16 buff_size, enum i40e_admin_queue_opc opcode, struct i40e_asq_cmd_details *cmd_details) argument 4229 i40e_aq_debug_dump(struct i40e_hw *hw, u8 cluster_id, u8 table_id, u32 start_index, u16 buff_size, void *buff, u16 *ret_buff_size, u8 *ret_next_table, u32 *ret_next_index, struct i40e_asq_cmd_details *cmd_details) argument 5160 i40e_aq_write_ddp(struct i40e_hw *hw, void *buff, u16 buff_size, u32 track_id, u32 *error_offset, u32 *error_info, struct i40e_asq_cmd_details *cmd_details) argument 5203 i40e_aq_get_ddp_list(struct i40e_hw *hw, void *buff, u16 buff_size, u8 flags, struct i40e_asq_cmd_details *cmd_details) argument [all...] |
H A D | i40e_nvm.c | 1001 u32 buff_size = 0; local 1029 buff_size = max_t(u32, aq_data_len, le16_to_cpu(aq_desc->datalen)); 1030 if (buff_size) { 1051 buff_size, &cmd_details);
|
H A D | i40e_prototype.h | 27 void *buff, /* can be NULL */ u16 buff_size, 33 u16 buff_size, 38 void *buff, /* can be NULL */ u16 buff_size, 45 u16 buff_size, 216 void *buff, u16 buff_size, u16 *data_size, 227 u8 mib_type, void *buff, u16 buff_size, 232 u8 mib_type, void *buff, u16 buff_size, 250 void *buff, u16 buff_size, 420 u8 table_id, u32 start_index, u16 buff_size, 469 u16 buff_size, u3 [all...] |
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_adminq.c | 611 * @buff_size: size of buffer for indirect commands 620 u16 buff_size, 672 if (buff_size > hw->aq.asq_buf_size) { 676 buff_size); 714 memcpy(dma_buff->va, buff, buff_size); 715 desc_on_ring->datalen = cpu_to_le16(buff_size); 729 buff, buff_size); 757 memcpy(buff, dma_buff->va, buff_size); 780 iavf_debug_aq(hw, IAVF_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); 617 iavf_asq_send_command(struct iavf_hw *hw, struct iavf_aq_desc *desc, void *buff, u16 buff_size, struct iavf_asq_cmd_details *cmd_details) argument
|
H A D | iavf_prototype.h | 28 u16 buff_size,
|