/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_gen4_pm_debugfs.c | 113 u32 *pm_info_regs, size_t buff_size, int table_len, 126 wr += scnprintf(&buff[wr], buff_size - wr, "%s: %#x\n", key, 135 u32 *pm_info_regs, size_t buff_size, 138 return pm_scnprint_table(buff, table, pm_info_regs, buff_size, 143 u32 *pm_info_regs, size_t buff_size, 146 return pm_scnprint_table(buff, table, pm_info_regs, buff_size, 112 pm_scnprint_table(char *buff, struct pm_status_row *table, u32 *pm_info_regs, size_t buff_size, int table_len, bool lowercase) argument 134 pm_scnprint_table_upper_keys(char *buff, struct pm_status_row *table, u32 *pm_info_regs, size_t buff_size, int table_len) argument 142 pm_scnprint_table_lower_keys(char *buff, struct pm_status_row *table, u32 *pm_info_regs, size_t buff_size, int table_len) argument
|
H A D | adf_admin.h | 24 int adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size);
|
/linux-master/sound/soc/stm/ |
H A D | stm32_adfsdm.c | 174 size_t buff_size = snd_pcm_lib_buffer_bytes(priv->substream); local 186 if ((priv->pos + src_size) > buff_size) { 189 buff_size - priv->pos); 192 buff_size - priv->pos); 193 cur_size -= buff_size - priv->pos; 204 priv->pos = (priv->pos + cur_size) % buff_size;
|
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_prototype.h | 28 u16 buff_size,
|
H A D | iavf_adminq.c | 611 * @buff_size: size of buffer for indirect commands 620 u16 buff_size, 672 if (buff_size > hw->aq.asq_buf_size) { 676 buff_size); 714 memcpy(dma_buff->va, buff, buff_size); 715 desc_on_ring->datalen = cpu_to_le16(buff_size); 729 buff, buff_size); 757 memcpy(buff, dma_buff->va, buff_size); 780 iavf_debug_aq(hw, IAVF_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); 617 iavf_asq_send_command(struct iavf_hw *hw, struct iavf_aq_desc *desc, void *buff, u16 buff_size, struct iavf_asq_cmd_details *cmd_details) argument
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_adminq.c | 744 * @buff_size: size of buffer for indirect commands 755 u16 buff_size, 806 if (buff_size > hw->aq.asq_buf_size) { 810 buff_size); 848 memcpy(dma_buff->va, buff, buff_size); 849 desc_on_ring->datalen = cpu_to_le16(buff_size); 863 buff, buff_size); 896 memcpy(buff, dma_buff->va, buff_size); 919 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); 948 * @buff_size 752 i40e_asq_send_command_atomic_exec(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, bool is_atomic_context) argument 956 i40e_asq_send_command_atomic(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, bool is_atomic_context) argument 975 i40e_asq_send_command(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details) argument 998 i40e_asq_send_command_atomic_v2(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, bool is_atomic_context, enum i40e_admin_queue_err *aq_status) argument 1020 i40e_asq_send_command_v2(struct i40e_hw *hw, struct i40e_aq_desc *desc, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details, enum i40e_admin_queue_err *aq_status) argument [all...] |
H A D | i40e_prototype.h | 27 void *buff, /* can be NULL */ u16 buff_size, 33 u16 buff_size, 38 void *buff, /* can be NULL */ u16 buff_size, 45 u16 buff_size, 216 void *buff, u16 buff_size, u16 *data_size, 227 u8 mib_type, void *buff, u16 buff_size, 232 u8 mib_type, void *buff, u16 buff_size, 250 void *buff, u16 buff_size, 427 u8 table_id, u32 start_index, u16 buff_size, 476 u16 buff_size, u3 [all...] |
H A D | i40e_common.c | 3334 * @buff_size: Size of the virtual buffer 3342 void *buff, u16 buff_size, u16 *data_size, 3361 if (buff_size > I40E_AQ_LARGE_BUF) 3364 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); 3476 * @buff_size: size of the buffer (in bytes) 3484 u8 mib_type, void *buff, u16 buff_size, 3495 if (buff_size == 0 || !buff) 3505 desc.datalen = cpu_to_le16(buff_size); 3508 if (buff_size > I40E_AQ_LARGE_BUF) 3511 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_detail 3341 i40e_aq_discover_capabilities(struct i40e_hw *hw, void *buff, u16 buff_size, u16 *data_size, enum i40e_admin_queue_opc list_type_opc, struct i40e_asq_cmd_details *cmd_details) argument 3483 i40e_aq_get_lldp_mib(struct i40e_hw *hw, u8 bridge_type, u8 mib_type, void *buff, u16 buff_size, u16 *local_len, u16 *remote_len, struct i40e_asq_cmd_details *cmd_details) argument 3533 i40e_aq_set_lldp_mib(struct i40e_hw *hw, u8 mib_type, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details) argument 3737 i40e_aq_get_cee_dcb_config(struct i40e_hw *hw, void *buff, u16 buff_size, struct i40e_asq_cmd_details *cmd_details) argument 3878 i40e_aq_tx_sched_cmd(struct i40e_hw *hw, u16 seid, void *buff, u16 buff_size, enum i40e_admin_queue_opc opcode, struct i40e_asq_cmd_details *cmd_details) argument 4482 i40e_aq_debug_dump(struct i40e_hw *hw, u8 cluster_id, u8 table_id, u32 start_index, u16 buff_size, void *buff, u16 *ret_buff_size, u8 *ret_next_table, u32 *ret_next_index, struct i40e_asq_cmd_details *cmd_details) argument 5413 i40e_aq_write_ddp(struct i40e_hw *hw, void *buff, u16 buff_size, u32 track_id, u32 *error_offset, u32 *error_info, struct i40e_asq_cmd_details *cmd_details) argument 5456 i40e_aq_get_ddp_list(struct i40e_hw *hw, void *buff, u16 buff_size, u8 flags, struct i40e_asq_cmd_details *cmd_details) argument [all...] |
H A D | i40e_nvm.c | 1379 u32 buff_size = 0; local 1407 buff_size = max_t(u32, aq_data_len, le16_to_cpu(aq_desc->datalen)); 1408 if (buff_size) { 1429 buff_size, &cmd_details);
|
/linux-master/drivers/firmware/efi/libstub/ |
H A D | mem.c | 54 m->buff_size = m->map_size = size;
|
/linux-master/drivers/net/ethernet/ibm/ |
H A D | ibmveth.c | 154 u32 buff_size, u32 pool_active) 158 pool->buff_size = buff_size; 229 skb = netdev_alloc_skb(adapter->netdev, pool->buff_size); 248 pool->buff_size, DMA_FROM_DEVICE); 260 desc.fields.flags_len = IBMVETH_BUF_VALID | pool->buff_size; 264 unsigned int len = min(pool->buff_size, 293 pool->dma_addr[index], pool->buff_size, 347 pool->buff_size, 386 adapter->rx_buff_pool[pool].buff_size, 152 ibmveth_init_buffer_pool(struct ibmveth_buff_pool *pool, u32 pool_index, u32 pool_size, u32 buff_size, u32 pool_active) argument [all...] |
H A D | ibmveth.h | 115 u32 buff_size; member in struct:ibmveth_buff_pool
|
H A D | ibmvnic.c | 571 * @buff_size: Size of each buffer in the LTB 573 * Allocate a set of LTBs to accommodate @num_buffs buffers of @buff_size 584 int buff_size) 597 dev_dbg(dev, "%s() num_buffs %d, buff_size %d\n", __func__, num_buffs, 598 buff_size); 600 ltb_size = rounddown(IBMVNIC_ONE_LTB_SIZE, buff_size); 601 tot_size = num_buffs * buff_size; 703 nbufs = ltb->size / rxpool->buff_size; 710 *offset = bufidx * rxpool->buff_size; 797 pool->buff_size); 582 alloc_ltb_set(struct ibmvnic_adapter *adapter, struct ibmvnic_ltb_set *ltb_set, int num_buffs, int buff_size) argument 1060 u64 buff_size; local 1300 u64 buff_size; local [all...] |
/linux-master/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
H A D | hw_atl_a0.c | 128 u32 buff_size = 0U; local 148 buff_size = HW_ATL_A0_TXBUF_MAX; 150 hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, buff_size, tc); 152 (buff_size * 156 (buff_size * 163 buff_size = HW_ATL_A0_RXBUF_MAX; 165 hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, buff_size, tc); 167 (buff_size * 171 (buff_size *
|
/linux-master/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_nic.c | 615 unsigned int buff_size = 0U; local 620 buff_size = AQ_CFG_TX_FRAME_MAX; 622 buff_size = frag_len; 625 buff_size, DMA_TO_DEVICE); 634 dx_buff->len = buff_size; 639 frag_len -= buff_size; 640 buff_offset += buff_size; 759 unsigned int buff_size = 0U; local 767 buff_size = AQ_CFG_TX_FRAME_MAX; 769 buff_size [all...] |
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_iwarp.h | 35 u32 buff_size; member in struct:qed_iwarp_ll2_buff
|
H A D | qed_iwarp.c | 1584 (u16)buf->buff_size, buf, 1); 1589 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buf->buff_size, 1903 if ((fpdu->mpa_frag_len + tcp_payload_size) > (u16)buf->buff_size) { 1905 "MPA ALIGN: Unexpected: buffer is not large enough for split fpdu buff_size = %d mpa_frag_len = %d, tcp_payload_size = %d, incomplete_bytes = %d\n", 1906 buf->buff_size, fpdu->mpa_frag_len, 1941 "MPA ALIGN: split fpdu buff_size = %d mpa_frag_len = %d, tcp_payload_size = %d, incomplete_bytes = %d\n", 1942 buf->buff_size, fpdu->mpa_frag_len, tcp_payload_size, 2474 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size, 2517 buffer->piggy_buf->buff_size, 2524 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size, 2597 qed_iwarp_ll2_alloc_buffers(struct qed_hwfn *p_hwfn, int num_rx_bufs, int buff_size, u8 ll2_handle) argument 2642 u32 buff_size; local [all...] |
/linux-master/drivers/usb/renesas_usbhs/ |
H A D | pipe.c | 474 u16 buff_size; local 489 buff_size = pipe_config->bufsize; 492 /* change buff_size to register value */ 493 bufnmb_cnt = (buff_size / 64) - 1; 495 dev_dbg(dev, "pipe : %d : buff_size 0x%x: bufnmb 0x%x\n", 496 pipe_num, buff_size, bufnmb);
|
/linux-master/drivers/dma/ |
H A D | mv_xor_v2.c | 100 * @buff_size: amount of bytes to be processed 129 u32 buff_size; member in struct:mv_xor_v2_descriptor 421 hw_descriptor->buff_size = len; 480 hw_descriptor->buff_size = len;
|
/linux-master/drivers/gpu/drm/bridge/cadence/ |
H A D | cdns-mhdp8546-hdcp.c | 83 u8 *buff, u16 buff_size) 88 for (i = 0; i < buff_size; i++) { 82 cdns_mhdp_secure_mailbox_recv_data(struct cdns_mhdp_device *mhdp, u8 *buff, u16 buff_size) argument
|
/linux-master/sound/soc/sof/ |
H A D | debug.c | 217 static int memory_info_update(struct snd_sof_dev *sdev, char *buf, size_t buff_size) argument 255 ret = scnprintf(buf + len, buff_size - len, "zone %d.%d used %#8x free %#8x\n",
|
/linux-master/drivers/ntb/ |
H A D | ntb_transport.c | 209 size_t buff_size; member in struct:ntb_transport_mw 803 mw->buff_size = 0; 832 if (mw->alloc_size > mw->buff_size) { 858 size_t xlat_size, buff_size; local 872 buff_size = round_up(size, xlat_align); 878 if (mw->buff_size) 883 mw->buff_size = buff_size; 884 mw->alloc_size = buff_size; 894 mw->buff_size [all...] |
/linux-master/drivers/s390/net/ |
H A D | lcs.h | 219 __u16 buff_size; member in struct:lcs_cmd::__anon2622::__anon2624
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_qp.c | 651 hr_qp->buff_size = 0; 661 hr_qp->buff_size += buf_size; 665 hr_qp->sge.offset = hr_qp->buff_size; 672 hr_qp->buff_size += buf_size; 676 hr_qp->rq.offset = hr_qp->buff_size; 683 hr_qp->buff_size += buf_size; 686 if (hr_qp->buff_size < 1)
|
/linux-master/drivers/crypto/ccree/ |
H A D | cc_buffer_mgr.c | 117 u32 buff_size, u32 *curr_nents, 124 new_nents = (*curr_nents + buff_size / CC_MAX_MLLI_ENTRY_SIZE + 1); 132 while (buff_size > CC_MAX_MLLI_ENTRY_SIZE) { 139 buff_size -= CC_MAX_MLLI_ENTRY_SIZE; 145 cc_lli_set_size(mlli_entry_p, buff_size); 116 cc_render_buff_to_mlli(struct device *dev, dma_addr_t buff_dma, u32 buff_size, u32 *curr_nents, u32 **mlli_entry_pp) argument
|