/linux-master/kernel/rcu/ |
H A D | tree.h | 484 static void rcu_nocb_gp_cleanup(struct swait_queue_head *sq);
|
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma.h | 106 struct erdma_cmdq_sq sq; member in struct:erdma_cmdq
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 347 idx = tail & (qp->sq.wqe_cnt - 1); 351 tail = qp->sq.w_list[idx].next; 353 tail = qp->sq.w_list[idx].next; 354 qp->sq.last_poll = tail; 402 wq = (is_send) ? &qp->sq : &qp->rq; 502 wq = &(*cur_qp)->sq; 534 wq = &(*cur_qp)->sq;
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | uk.c | 90 * irdma_clr_wqes - clear next 128 sq entries 96 struct irdma_qp_quanta *sq; local 101 sq = qp->sq_base + wqe_idx; 103 memset(sq, qp->swqe_polarity ? 0 : 0xFF, 104 128 * sizeof(*sq)); 106 memset(sq, qp->swqe_polarity ? 0xFF : 0, 107 128 * sizeof(*sq)); 229 * @info: post sq information 230 * @post_sq: flag to post sq 323 * @info: post sq informatio [all...] |
H A D | user.h | 364 struct irdma_qp_quanta *sq; member in struct:irdma_qp_uk_init_info
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | vcn_v4_0_5.c | 149 fw_shared->sq.is_enabled = 1; 195 fw_shared->sq.is_enabled = 0; 928 fw_shared->sq.queue_mode |= FW_QUEUE_RING_RESET; 939 fw_shared->sq.queue_mode &= ~(FW_QUEUE_RING_RESET | FW_QUEUE_DPG_HOLD_OFF); 1112 fw_shared->sq.queue_mode |= FW_QUEUE_RING_RESET; 1123 fw_shared->sq.queue_mode &= ~(FW_QUEUE_RING_RESET | FW_QUEUE_DPG_HOLD_OFF); 1172 fw_shared->sq.queue_mode |= FW_QUEUE_DPG_HOLD_OFF;
|
H A D | vcn_v4_0_3.c | 136 fw_shared->sq.is_enabled = true; 180 fw_shared->sq.is_enabled = cpu_to_le32(false); 830 fw_shared->sq.queue_mode |= FW_QUEUE_RING_RESET; 840 fw_shared->sq.queue_mode &= ~(FW_QUEUE_RING_RESET | FW_QUEUE_DPG_HOLD_OFF); 843 fw_shared->sq.queue_mode &= cpu_to_le32(~FW_QUEUE_RING_RESET); 1204 fw_shared->sq.queue_mode &= 1260 fw_shared->sq.queue_mode |= FW_QUEUE_DPG_HOLD_OFF;
|
H A D | vcn_v4_0.c | 109 fw_shared->sq.is_enabled = 1; 227 fw_shared->sq.is_enabled = 0; 1017 fw_shared->sq.queue_mode |= FW_QUEUE_RING_RESET; 1028 fw_shared->sq.queue_mode &= ~(FW_QUEUE_RING_RESET | FW_QUEUE_DPG_HOLD_OFF); 1200 fw_shared->sq.queue_mode |= FW_QUEUE_RING_RESET; 1211 fw_shared->sq.queue_mode &= ~(FW_QUEUE_RING_RESET | FW_QUEUE_DPG_HOLD_OFF); 1510 fw_shared->sq.queue_mode |= FW_QUEUE_DPG_HOLD_OFF;
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 202 if (msg_len > qp->sq.ext_sge_cnt * HNS_ROCE_SGE_SIZE) { 625 hr_reg_write(&sq_db, DB_PI, qp->sq.head); 628 hns_roce_write64(hr_dev, (__le32 *)&sq_db, qp->sq.db_reg); 679 hr_reg_write(rc_sq_wqe, RC_SEND_WQE_WQE_INDEX, qp->sq.head); 681 hns_roce_write512(hr_dev, wqe, qp->sq.db_reg); 699 spin_lock_irqsave(&qp->sq.lock, flags); 711 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { 717 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); 719 if (unlikely(wr->num_sge > qp->sq [all...] |
H A D | hns_roce_main.c | 1034 spin_lock_irqsave(&hr_qp->sq.lock, flags_qp); 1035 if (hr_qp->sq.tail != hr_qp->sq.head) 1037 spin_unlock_irqrestore(&hr_qp->sq.lock, flags_qp);
|
/linux-master/drivers/nvme/target/ |
H A D | zns.c | 71 u8 zasl = req->sq->ctrl->subsys->zasl; 72 struct nvmet_ctrl *ctrl = req->sq->ctrl;
|
/linux-master/drivers/staging/ks7010/ |
H A D | ks_wlan.h | 205 u8 sq; member in struct:local_ap
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 496 qp->sq.db_index = ucmd.sq_db_index; 543 init_attr->cap.max_send_wr = qp->sq.max; 545 init_attr->cap.max_send_sge = qp->sq.max_gs; 564 to_mqp(qp)->sq.db_index);
|
/linux-master/drivers/net/wireless/ath/ath11k/ |
H A D | qmi.h | 130 struct sockaddr_qrtr sq; member in struct:ath11k_qmi
|
H A D | qmi.c | 3011 struct sockaddr_qrtr *sq, 3057 struct sockaddr_qrtr *sq, 3069 struct sockaddr_qrtr *sq, 3087 struct sockaddr_qrtr *sq, 3101 struct sockaddr_qrtr *sq, 3161 struct sockaddr_qrtr *sq = &qmi->sq; local 3164 sq->sq_family = AF_QIPCRTR; 3165 sq->sq_node = service->node; 3166 sq 3010 ath11k_qmi_msg_mem_request_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *data) argument 3056 ath11k_qmi_msg_mem_ready_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *decoded) argument 3068 ath11k_qmi_msg_fw_ready_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *decoded) argument 3086 ath11k_qmi_msg_cold_boot_cal_done_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *decoded) argument 3100 ath11k_qmi_msg_fw_init_done_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *decoded) argument [all...] |
/linux-master/include/linux/ |
H A D | io_uring_types.h | 120 * The kernel controls head of the sq ring and the tail of the cq ring, 121 * and the application controls tail of the sq ring and the head of the 124 struct io_uring sq, cq; member in struct:io_rings 152 * for IORING_SQ_NEED_WAKEUP after updating the sq tail. 361 struct io_sq_data *sq_data; /* if using sq thread polling */
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_verbs.h | 215 struct rxe_sq sq; member in struct:rxe_qp
|
/linux-master/drivers/iommu/intel/ |
H A D | irq_remapping.c | 284 unsigned int sq, unsigned int sid) 289 irte->sq = sq; 1360 irte->sid, irte->sq, irte->svt); 283 set_irte_sid(struct irte *irte, unsigned int svt, unsigned int sq, unsigned int sid) argument
|
/linux-master/net/qrtr/ |
H A D | af_qrtr.c | 1150 struct sockaddr_qrtr *sq; local 1177 sq = (struct sockaddr_qrtr *)&ifr.ifr_addr; 1178 *sq = ipc->us;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ipoib/ |
H A D | ipoib.c | 131 struct mlx5e_sq_stats *sq_stats = &channel_stats->sq[j]; 707 struct mlx5e_txqsq *sq = epriv->txq2sq[skb_get_queue_mapping(skb)]; local 711 mlx5i_sq_xmit(sq, skb, &mah->av, dqpn, ipriv->qkey, netdev_xmit_more());
|
/linux-master/drivers/scsi/qedi/ |
H A D | qedi_fw.c | 1042 task_params.sqe = &ep->sq[sq_idx]; 1116 task_params.sqe = &ep->sq[sq_idx]; 1490 task_params.sqe = &ep->sq[sq_idx]; 1614 task_params.sqe = &ep->sq[sq_idx]; 1731 task_params.sqe = &ep->sq[sq_idx]; 2093 task_params.sqe = &ep->sq[sq_idx]; 2148 task_params.sqe = &ep->sq[sq_idx];
|
/linux-master/drivers/target/ |
H A D | target_core_transport.c | 1882 struct se_cmd_queue *sq = container_of(work, struct se_cmd_queue, work); local 1888 cmd_list = llist_del_all(&sq->cmd_list); 1915 struct se_cmd_queue *sq; local 1917 sq = &se_dev->queues[cpu].sq; 1918 llist_add(&se_cmd->se_cmd_list, &sq->cmd_list); 1919 queue_work_on(cpu, target_submission_wq, &sq->work);
|
/linux-master/drivers/net/wireless/ath/ath12k/ |
H A D | qmi.c | 3157 struct sockaddr_qrtr *sq, 3193 struct sockaddr_qrtr *sq, 3205 struct sockaddr_qrtr *sq, 3248 struct sockaddr_qrtr *sq = &qmi->sq; local 3251 sq->sq_family = AF_QIPCRTR; 3252 sq->sq_node = service->node; 3253 sq->sq_port = service->port; 3255 ret = kernel_connect(qmi_hdl->sock, (struct sockaddr *)sq, 3256 sizeof(*sq), 3156 ath12k_qmi_msg_mem_request_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *data) argument 3192 ath12k_qmi_msg_mem_ready_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *decoded) argument 3204 ath12k_qmi_msg_fw_ready_cb(struct qmi_handle *qmi_hdl, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *decoded) argument [all...] |
/linux-master/drivers/crypto/hisilicon/sec2/ |
H A D | sec_crypto.c | 1054 struct skcipher_request *sq = req->c_req.sk_req; local 1056 return sec_cipher_map(ctx, req, sq->src, sq->dst); 1061 struct skcipher_request *sq = req->c_req.sk_req; local 1063 sec_cipher_unmap(ctx, req, sq->src, sq->dst);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_ethtool.c | 624 coal->use_adaptive_tx_coalesce = !!c->sq[0].dim; 630 cur_moder = net_dim_get_tx_moderation(c->sq[0].dim->mode, 631 c->sq[0].dim->profile_ix); 672 mlx5e_modify_cq_moderation(mdev, &c->sq[tc].cq.mcq, 857 err = mlx5e_dim_tx_change(&c->sq[tc], tx_dim_enabled); 876 mlx5_core_modify_cq_moderation(mdev, &c->sq[tc].cq.mcq,
|