/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/sec2/ |
H A D | ga102.c | 44 for (i = 0; i < ARRAY_SIZE(msg.queue_info); i++) { 45 if (msg.queue_info[i].id == NV_SEC2_INIT_MSG_QUEUE_ID_MSGQ) { 46 nvkm_falcon_msgq_init(sec2->msgq, msg.queue_info[i].index, 47 msg.queue_info[i].offset, 48 msg.queue_info[i].size); 50 nvkm_falcon_cmdq_init(sec2->cmdq, msg.queue_info[i].index, 51 msg.queue_info[i].offset, 52 msg.queue_info[i].size);
|
H A D | gp102.c | 135 for (i = 0; i < ARRAY_SIZE(msg.queue_info); i++) { 136 if (msg.queue_info[i].id == NV_SEC2_INIT_MSG_QUEUE_ID_MSGQ) { 138 msg.queue_info[i].index, 139 msg.queue_info[i].offset, 140 msg.queue_info[i].size); 143 msg.queue_info[i].index, 144 msg.queue_info[i].offset, 145 msg.queue_info[i].size);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
H A D | gm20b.c | 172 nvkm_falcon_cmdq_init(pmu->hpq, msg.queue_info[0].index, 173 msg.queue_info[0].offset, 174 msg.queue_info[0].size); 175 nvkm_falcon_cmdq_init(pmu->lpq, msg.queue_info[1].index, 176 msg.queue_info[1].offset, 177 msg.queue_info[1].size); 178 nvkm_falcon_msgq_init(pmu->msgq, msg.queue_info[4].index, 179 msg.queue_info[4].offset, 180 msg.queue_info[4].size);
|
/linux-master/drivers/gpu/drm/nouveau/include/nvfw/ |
H A D | sec2.h | 31 } queue_info[2]; member in struct:nv_sec2_init_msg 52 } queue_info[2]; member in struct:nv_sec2_init_msg_v1
|
H A D | pmu.h | 40 } queue_info[5]; member in struct:nv_pmu_init_msg
|
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | sta.c | 316 sta_id = mvm->queue_info[queue].ra_sta_id; 390 if (WARN_ON(mvm->queue_info[queue].tid_bitmap == 0)) 393 mvm->queue_info[queue].tid_bitmap &= ~BIT(tid); 395 cmd.action = mvm->queue_info[queue].tid_bitmap ? 398 mvm->queue_info[queue].status = IWL_MVM_QUEUE_FREE; 403 mvm->queue_info[queue].tid_bitmap); 409 cmd.sta_id = mvm->queue_info[queue].ra_sta_id; 410 cmd.tid = mvm->queue_info[queue].txq_tid; 413 WARN(mvm->queue_info[queue].tid_bitmap, 415 queue, mvm->queue_info[queu [all...] |
H A D | tx.c | 1124 unsigned long queue_tid_bitmap = mvm->queue_info[txq_id].tid_bitmap; 1132 if (time_before(mvm->queue_info[txq_id].last_frame_time[tid] + 1276 mvm->queue_info[txq_id].last_frame_time[tid] = jiffies; 1288 if (unlikely(mvm->queue_info[txq_id].status == 2383 struct iwl_flush_queue_info *queue_info = &rsp->queues[i]; local 2384 int tid = le16_to_cpu(queue_info->tid); 2385 int read_before = le16_to_cpu(queue_info->read_before_flush); 2386 int read_after = le16_to_cpu(queue_info->read_after_flush); 2387 int queue_num = le16_to_cpu(queue_info->queue_num);
|
H A D | fw.c | 457 memset(&mvm->queue_info, 0, sizeof(mvm->queue_info)); 464 mvm->queue_info[IWL_MVM_DQA_CMD_QUEUE].tid_bitmap =
|
H A D | ops.c | 1766 mvm->queue_info[hw_queue].ra_sta_id; 1792 tid_bitmap = mvm->queue_info[hw_queue].tid_bitmap;
|
H A D | mvm.h | 894 struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES]; member in union:iwl_mvm::__anon55
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 273 static int offload_tso(struct hinic_sq_task *task, u32 *queue_info, argument 347 hinic_set_tso_inner_l4(task, queue_info, l4_offload, l4_len, offset, 353 static int offload_csum(struct hinic_sq_task *task, u32 *queue_info, argument 434 hinic_set_cs_inner_l4(task, queue_info, l4_offload, l4_len, offset); 439 static void offload_vlan(struct hinic_sq_task *task, u32 *queue_info, argument 445 *queue_info |= HINIC_SQ_CTRL_SET(vlan_pri, QUEUE_INFO_PRI); 449 u32 *queue_info) 455 enabled = offload_tso(task, queue_info, skb); 459 enabled = offload_csum(task, queue_info, skb); 468 offload_vlan(task, queue_info, vlan_ta 448 hinic_tx_offload(struct sk_buff *skb, struct hinic_sq_task *task, u32 *queue_info) argument [all...] |
H A D | hinic_hw_qp.c | 489 ctrl->queue_info = HINIC_SQ_CTRL_SET(HINIC_MSS_DEFAULT, 536 void hinic_set_cs_inner_l4(struct hinic_sq_task *task, u32 *queue_info, argument 552 *queue_info |= HINIC_SQ_CTRL_SET(offset, QUEUE_INFO_PLDOFF) | 556 *queue_info = HINIC_SQ_CTRL_CLEAR(*queue_info, QUEUE_INFO_MSS); 557 *queue_info |= HINIC_SQ_CTRL_SET(mss, QUEUE_INFO_MSS); 560 void hinic_set_tso_inner_l4(struct hinic_sq_task *task, u32 *queue_info, argument 577 *queue_info |= HINIC_SQ_CTRL_SET(offset, QUEUE_INFO_PLDOFF) | 583 *queue_info = HINIC_SQ_CTRL_CLEAR(*queue_info, QUEUE_INFO_MS [all...] |
H A D | hinic_hw_qp.h | 168 u32 *queue_info, 173 u32 *queue_info,
|
H A D | hinic_hw_wqe.h | 365 u32 queue_info; member in struct:hinic_sq_ctrl
|
/linux-master/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 135 * @queue_info: The &struct ath5k_txq_info to fill 139 struct ath5k_txq_info *queue_info) 141 memcpy(queue_info, &ah->ah_txq[queue], sizeof(struct ath5k_txq_info)); 198 * @queue_info: The &struct ath5k_txq_info to use 204 struct ath5k_txq_info *queue_info) 228 queue = queue_info->tqi_subtype; 250 if (queue_info != NULL) { 251 queue_info->tqi_type = queue_type; 252 ret = ath5k_hw_set_tx_queueprops(ah, queue, queue_info); 138 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument 203 ath5k_hw_setup_tx_queue(struct ath5k_hw *ah, enum ath5k_tx_queue queue_type, struct ath5k_txq_info *queue_info) argument
|
H A D | ath5k.h | 1559 struct ath5k_txq_info *queue_info); 1561 const struct ath5k_txq_info *queue_info); 1564 struct ath5k_txq_info *queue_info);
|
/linux-master/drivers/media/platform/chips-media/wave5/ |
H A D | wave5-vpuapi.c | 628 struct queue_status_info *queue_info = parameter; local 630 queue_info->instance_queue_count = p_dec_info->instance_queue_count; 631 queue_info->report_queue_count = p_dec_info->report_queue_count; 906 struct queue_status_info *queue_info = parameter; local 908 queue_info->instance_queue_count = p_enc_info->instance_queue_count; 909 queue_info->report_queue_count = p_enc_info->report_queue_count;
|
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_mbx.c | 516 struct hclge_mbx_vf_queue_info *queue_info; local 520 queue_info = (struct hclge_mbx_vf_queue_info *)resp_msg->data; 521 queue_info->num_tqps = cpu_to_le16(vport->alloc_tqps); 522 queue_info->rss_size = cpu_to_le16(vport->nic.kinfo.rss_size); 523 queue_info->rx_buf_len = cpu_to_le16(hdev->rx_buf_len);
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_lag.c | 404 qbuf->queue_info[count].q_handle = cpu_to_le16(qid); 405 qbuf->queue_info[count].tc = tc; 406 qbuf->queue_info[count].q_teid = cpu_to_le32(q_ctx->q_teid); 539 qbuf_size = struct_size(qbuf, queue_info, numq); 888 qbuf_size = struct_size(qbuf, queue_info, numq); 1909 qbuf_size = struct_size(qbuf, queue_info, numq);
|
H A D | ice_adminq_cmd.h | 2066 struct ice_aqc_cfg_txq_perq queue_info[]; member in struct:ice_aqc_cfg_txqs_buf
|
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3vf/ |
H A D | hclgevf_main.c | 175 struct hclge_mbx_vf_queue_info *queue_info; local 190 queue_info = (struct hclge_mbx_vf_queue_info *)resp_msg; 191 hdev->num_tqps = le16_to_cpu(queue_info->num_tqps); 192 hdev->rss_size_max = le16_to_cpu(queue_info->rss_size); 193 hdev->rx_buf_len = le16_to_cpu(queue_info->rx_buf_len);
|
/linux-master/sound/core/seq/ |
H A D | seq_clientmgr.c | 2282 struct snd_seq_queue_info queue_info; member in union:__anon4075
|