/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | tx.c | 46 static void mlx5e_xsk_tx_post_err(struct mlx5e_xdpsq *sq, argument 49 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); 50 struct mlx5e_xdp_wqe_info *wi = &sq->db.wqe_info[pi]; 56 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); 57 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, *xdpi); 58 if (xp_tx_metadata_enabled(sq->xsk_pool)) 59 mlx5e_xdpi_fifo_push(&sq 64 mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigned int budget) argument [all...] |
H A D | tx.h | 13 bool mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigned int budget);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_tx.c | 46 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) argument 52 mlx5e_dma_get(sq, --sq->dma_fifo_pc); 54 mlx5e_tx_dma_unmap(sq->pdev, last_pushed_dma); 119 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, argument 123 if (unlikely(mlx5e_ipsec_txwqe_build_eseg_csum(sq, skb, eseg))) 131 sq->stats->csum_partial_inner++; 134 sq->stats->csum_partial++; 139 sq->stats->csum_partial++; 142 sq 149 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb, int *hopbyhop) argument 177 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, unsigned char *skb_data, u16 headlen, struct mlx5_wqe_data_seg *dseg) argument 243 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel) argument 262 mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5e_tx_attr *attr) argument 342 mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) argument 350 mlx5e_tx_flush(struct mlx5e_txqsq *sq) argument 371 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, u8 num_dma, struct mlx5e_tx_wqe_info *wi, struct mlx5_wqe_ctrl_seg *cseg, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) argument 423 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) argument 511 mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) argument 519 mlx5e_tx_mpwqe_session_start(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) argument 543 mlx5e_tx_mpwqe_session_is_active(struct mlx5e_txqsq *sq) argument 548 mlx5e_tx_mpwqe_add_dseg(struct mlx5e_txqsq *sq, struct mlx5e_xmit_data *txd) argument 566 mlx5e_tx_mpwqe_session_complete(struct mlx5e_txqsq *sq) argument 598 mlx5e_sq_xmit_mpwqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) argument 647 mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq) argument 662 mlx5e_txwqe_build_eseg(struct mlx5e_priv *priv, struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5_wqe_eth_seg *eseg, u16 ihs) argument 679 struct mlx5e_txqsq *sq; local 732 mlx5e_tx_wi_dma_unmap(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) argument 744 mlx5e_consume_skb(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_cqe64 *cqe, int napi_budget) argument 762 mlx5e_tx_wi_consume_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, struct mlx5_cqe64 *cqe, int napi_budget) argument 774 mlx5e_txqsq_wake(struct mlx5e_txqsq *sq) argument 788 struct mlx5e_txqsq *sq; local 889 mlx5e_tx_wi_kfree_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi) argument 897 mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq) argument 972 mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_av *av, u32 dqpn, u32 dqkey, bool xmit_more) argument [all...] |
H A D | en_txrx.c | 49 static void mlx5e_handle_tx_dim(struct mlx5e_txqsq *sq) argument 51 struct mlx5e_sq_stats *stats = sq->stats; 54 if (unlikely(!test_bit(MLX5E_SQ_STATE_DIM, &sq->state))) 57 dim_update_sample(sq->cq.event_ctr, stats->packets, stats->bytes, &dim_sample); 58 net_dim(&sq->dim, dim_sample); 73 void mlx5e_trigger_irq(struct mlx5e_icosq *sq) argument 75 struct mlx5_wq_cyc *wq = &sq->wq; 77 u16 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); 79 sq->db.wqe_info[pi] = (struct mlx5e_icosq_wqe_info) { 84 nopwqe = mlx5e_post_nop(wq, sq 157 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); local 225 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); local [all...] |
/linux-master/drivers/nvme/target/ |
H A D | fabrics-cmd-auth.c | 17 struct nvmet_sq *sq = container_of(to_delayed_work(work), local 21 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid); 22 sq->dhchap_step = NVME_AUTH_DHCHAP_MESSAGE_NEGOTIATE; 23 sq->dhchap_tid = -1; 26 void nvmet_auth_sq_init(struct nvmet_sq *sq) argument 29 INIT_DELAYED_WORK(&sq->auth_expired_work, nvmet_auth_expired_work); 30 sq->authenticated = false; 31 sq [all...] |
H A D | auth.c | 222 void nvmet_auth_sq_free(struct nvmet_sq *sq) argument 224 cancel_delayed_work(&sq->auth_expired_work); 225 kfree(sq->dhchap_c1); 226 sq->dhchap_c1 = NULL; 227 kfree(sq->dhchap_c2); 228 sq->dhchap_c2 = NULL; 229 kfree(sq->dhchap_skey); 230 sq->dhchap_skey = NULL; 257 if (req->sq->ctrl->host_key && 258 !req->sq [all...] |
/linux-master/tools/include/io_uring/ |
H A D | mini_liburing.h | 55 struct io_uring_sq sq; member in struct:io_uring 69 struct io_uring_sq *sq, struct io_uring_cq *cq) 75 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned int); 76 ptr = mmap(0, sq->ring_sz, PROT_READ | PROT_WRITE, 80 sq->khead = ptr + p->sq_off.head; 81 sq->ktail = ptr + p->sq_off.tail; 82 sq->kring_mask = ptr + p->sq_off.ring_mask; 83 sq->kring_entries = ptr + p->sq_off.ring_entries; 84 sq->kflags = ptr + p->sq_off.flags; 85 sq 68 io_uring_mmap(int fd, struct io_uring_params *p, struct io_uring_sq *sq, struct io_uring_cq *cq) argument 154 struct io_uring_sq *sq = &ring->sq; local 187 struct io_uring_sq *sq = &ring->sq; local 222 struct io_uring_sq *sq = &ring->sq; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | xdp.c | 61 mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq, argument 96 dma_addr = dma_map_single(sq->pdev, xdptxd->data, xdptxd->len, 98 if (dma_mapping_error(sq->pdev, dma_addr)) { 105 if (unlikely(!INDIRECT_CALL_2(sq->xmit_xdp_frame, mlx5e_xmit_xdp_frame_mpwqe, 106 mlx5e_xmit_xdp_frame, sq, xdptxd, 0, NULL))) 110 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, 112 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, 114 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, 126 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd->len, DMA_BIDIRECTIONAL); 140 dma_sync_single_for_device(sq 349 mlx5e_xdpsq_get_next_pi(struct mlx5e_xdpsq *sq, u16 size) argument 378 mlx5e_xdp_mpwqe_session_start(struct mlx5e_xdpsq *sq) argument 400 mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq) argument 428 mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq) argument 450 mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result, struct xsk_tx_metadata *meta) argument 511 mlx5e_xmit_xdp_frame_check_stop_room(struct mlx5e_xdpsq *sq, int stop_room) argument 523 mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq) argument 529 mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result, struct xsk_tx_metadata *meta) argument 659 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, struct xdp_frame_bulk *bq, struct mlx5e_cq *cq, struct mlx5_cqe64 *cqe) argument 751 struct mlx5e_xdpsq *sq; local 819 mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq) argument 851 struct mlx5e_xdpsq *sq; local 968 mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw) argument [all...] |
H A D | reporter_tx.c | 22 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) argument 24 struct mlx5_core_dev *dev = sq->mdev; 30 if (sq->cc == sq->pc) 36 netdev_err(sq->netdev, 37 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n", 38 sq->sqn, sq->cc, sq 43 mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) argument 53 mlx5e_health_sq_put_sw_state(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq) argument 72 struct mlx5e_txqsq *sq; local 127 struct mlx5e_txqsq *sq; member in struct:mlx5e_tx_timeout_ctx 136 struct mlx5e_txqsq *sq; local 225 mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) argument 249 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) argument 341 struct mlx5e_txqsq *sq = &c->sq[tc]; local 366 struct mlx5e_txqsq *sq = ctx; local 434 struct mlx5e_txqsq *sq = &c->sq[tc]; local 442 struct mlx5e_txqsq *sq = &ptp_ch->ptpsq[tc].txqsq; local 470 mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq) argument 484 mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq) argument [all...] |
H A D | xdp.h | 108 void mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq); 110 void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq); 111 void mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw); 119 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, 123 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, 127 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq)); 128 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq)); 158 static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) argument 160 if (sq->doorbell_cseg) { 161 mlx5e_notify_hw(&sq 169 mlx5e_xdp_get_inline_state(struct mlx5e_xdpsq *sq, bool cur) argument 200 mlx5e_xdp_mpwqe_add_dseg(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdpsq_stats *stats) argument [all...] |
H A D | txrx.h | 61 void mlx5e_trigger_irq(struct mlx5e_icosq *sq); 95 void mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq); 119 #define MLX5E_TX_FETCH_WQE(sq, pi) \ 120 ((struct mlx5e_tx_wqe *)mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_tx_wqe))) 168 static inline u16 mlx5e_txqsq_get_next_pi(struct mlx5e_txqsq *sq, u16 size) argument 170 struct mlx5_wq_cyc *wq = &sq->wq; 173 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); 178 wi = &sq->db.wqe_info[pi]; 186 mlx5e_post_nop(wq, sq->sqn, &sq 230 mlx5e_icosq_get_next_pi(struct mlx5e_icosq *sq, u16 size) argument 285 mlx5e_dma_get(struct mlx5e_txqsq *sq, u32 i) argument 291 mlx5e_dma_push(struct mlx5e_txqsq *sq, dma_addr_t addr, u32 size, enum mlx5e_dma_map_type map_type) argument 496 mlx5e_icosq_can_post_wqe(struct mlx5e_icosq *sq, u16 wqe_size) argument [all...] |
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos_sq.c | 38 struct otx2_snd_queue *sq; local 76 sq = &qset->sq[qidx]; 77 sq->sqb_count = 0; 78 sq->sqb_ptrs = kcalloc(num_sqbs, sizeof(*sq->sqb_ptrs), GFP_KERNEL); 79 if (!sq->sqb_ptrs) { 89 sq->sqb_ptrs[sq->sqb_count++] = (u64)bufptr; 96 if (!sq 120 struct otx2_snd_queue *sq; local 263 struct otx2_snd_queue *sq; local [all...] |
H A D | otx2_txrx.c | 116 struct otx2_snd_queue *sq, 124 sg = &sq->sg[snd_comp->sqe_id]; 135 struct otx2_snd_queue *sq, 151 sg = &sq->sg[snd_comp->sqe_id]; 157 timestamp = ((u64 *)sq->timestamps->base)[snd_comp->sqe_id]; 451 struct otx2_snd_queue *sq; local 463 sq = &pfvf->qset.sq[qidx]; 476 otx2_xdp_snd_pkt_handler(pfvf, sq, cqe); 478 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[qid 115 otx2_xdp_snd_pkt_handler(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_cqe_tx_s *cqe) argument 133 otx2_snd_pkt_handler(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, struct otx2_snd_queue *sq, struct nix_cqe_tx_s *cqe, int budget, int *tx_pkts, int *tx_bytes) argument 595 otx2_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx) argument 614 otx2_sqe_add_sg(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int num_segs, int *offset) argument 659 otx2_sqe_add_ext(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int *offset) argument 727 otx2_sqe_add_mem(struct otx2_snd_queue *sq, int *offset, int alg, u64 iova, int ptp_offset, u64 base_ns, bool udp_csum_crt) argument 750 otx2_sqe_add_hdr(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_sqe_hdr_s *sqe_hdr, struct sk_buff *skb, u16 qidx) argument 798 otx2_dma_map_tso_skb(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int sqe, int hdr_len) argument 831 otx2_tso_frag_dma_addr(struct otx2_snd_queue *sq, struct sk_buff *skb, int seg, u64 seg_addr, int hdr_len, int sqe) argument 849 otx2_sqe_tso_add_sg(struct otx2_snd_queue *sq, struct sg_list *list, int *offset) argument 880 otx2_sq_append_tso(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) argument 1078 otx2_set_txtstamp(struct otx2_nic *pfvf, struct sk_buff *skb, struct otx2_snd_queue *sq, int *offset) argument 1141 otx2_sq_append_skb(struct net_device *netdev, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) argument 1252 struct otx2_snd_queue *sq; local 1319 struct otx2_snd_queue *sq; local 1347 otx2_xdp_sqe_add_sg(struct otx2_snd_queue *sq, u64 dma_addr, int len, int *offset) argument 1370 struct otx2_snd_queue *sq; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_utils.h | 62 #define MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi) \ 64 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_static_params_wqe))) 66 #define MLX5E_TLS_FETCH_SET_PROGRESS_PARAMS_WQE(sq, pi) \ 68 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_progress_params_wqe))) 70 #define MLX5E_TLS_FETCH_GET_PROGRESS_PARAMS_WQE(sq, pi) \ 72 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_get_tls_progress_params_wqe))) 74 #define MLX5E_TLS_FETCH_DUMP_WQE(sq, pi) \ 76 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_dump_wqe)))
|
H A D | ktls_txrx.h | 19 bool mlx5e_ktls_handle_tx_skb(struct net_device *netdev, struct mlx5e_txqsq *sq, 27 struct mlx5e_icosq *sq); 29 void mlx5e_ktls_tx_handle_resync_dump_comp(struct mlx5e_txqsq *sq, 33 mlx5e_ktls_tx_try_handle_resync_dump_comp(struct mlx5e_txqsq *sq, argument 38 mlx5e_ktls_tx_handle_resync_dump_comp(sq, wi, dma_fifo_cc); 60 mlx5e_ktls_tx_try_handle_resync_dump_comp(struct mlx5e_txqsq *sq, argument
|
H A D | ktls_rx.c | 129 static void icosq_fill_wi(struct mlx5e_icosq *sq, u16 pi, argument 132 sq->db.wqe_info[pi] = *wi; 136 post_static_params(struct mlx5e_icosq *sq, argument 144 if (unlikely(!mlx5e_icosq_can_post_wqe(sq, num_wqebbs))) 147 pi = mlx5e_icosq_get_next_pi(sq, num_wqebbs); 148 wqe = MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi); 149 mlx5e_ktls_build_static_params(wqe, sq->pc, sq->sqn, &priv_rx->crypto_info, 159 icosq_fill_wi(sq, pi, &wi); 160 sq 166 post_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx, u32 next_record_tcp_sn) argument 201 struct mlx5e_icosq *sq; local 252 resync_post_get_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx) argument 335 struct mlx5e_icosq *sq; local 367 struct mlx5e_icosq *sq; local 423 mlx5e_ktls_handle_get_psv_completion(struct mlx5e_icosq_wqe_info *wi, struct mlx5e_icosq *sq) argument 724 struct mlx5e_icosq *sq; local [all...] |
/linux-master/sound/oss/dmasound/ |
H A D | dmasound_core.c | 411 static int sq_allocate_buffers(struct sound_queue *sq, int num, int size) argument 415 if (sq->buffers) 417 sq->numBufs = num; 418 sq->bufSize = size; 419 sq->buffers = kmalloc_array (num, sizeof(char *), GFP_KERNEL); 420 if (!sq->buffers) 423 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); 424 if (!sq->buffers[i]) { 426 dmasound.mach.dma_free(sq->buffers[i], size); 427 kfree(sq 435 sq_release_buffers(struct sound_queue *sq) argument 448 sq_setup(struct sound_queue *sq) argument 688 sq_init_waitqueue(struct sound_queue *sq) argument 707 sq_open2(struct sound_queue *sq, struct file *file, fmode_t mode, int numbufs, int bufsize) argument 953 set_queue_frags(struct sound_queue *sq, int bufs, int size) argument [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | wr.h | 27 * @sq - SQ buffer. 33 static inline void *get_sq_edge(struct mlx5_ib_wq *sq, u32 idx) argument 38 (&sq->fbc, 39 mlx5_frag_buf_get_idx_last_contig_stride(&sq->fbc, idx)); 46 * @sq: SQ buffer. 51 static inline void handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, argument 59 idx = (sq->cur_post + (wqe_sz >> 2)) & (sq->wqe_cnt - 1); 60 *cur_edge = get_sq_edge(sq, idx); 62 *seg = mlx5_frag_buf_get_wqe(&sq 74 mlx5r_memcpy_send_wqe(struct mlx5_ib_wq *sq, void **cur_edge, void **seg, u32 *wqe_sz, const void *src, size_t n) argument [all...] |
H A D | ib_rep.h | 18 struct mlx5_ib_sq *sq, 27 struct mlx5_ib_sq *sq, 26 create_flow_rule_vport_sq(struct mlx5_ib_dev *dev, struct mlx5_ib_sq *sq, u32 port) argument
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 8 (qinfo)->sq.head = prefix##_ATQH; \ 9 (qinfo)->sq.tail = prefix##_ATQT; \ 10 (qinfo)->sq.len = prefix##_ATQLEN; \ 11 (qinfo)->sq.bah = prefix##_ATQBAH; \ 12 (qinfo)->sq.bal = prefix##_ATQBAL; \ 13 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \ 14 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \ 15 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \ 16 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \ 77 if (cq->sq 860 struct ice_ctl_q_ring *sq = &cq->sq; local [all...] |
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.c | 20 static inline void nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, 505 struct snd_queue *sq, int q_len, int qidx) 509 err = nicvf_alloc_q_desc_mem(nic, &sq->dmem, q_len, SND_QUEUE_DESC_SIZE, 514 sq->desc = sq->dmem.base; 515 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); 516 if (!sq->skbuff) 519 sq->head = 0; 520 sq->tail = 0; 521 sq 504 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) argument 554 nicvf_unmap_sndq_buffers(struct nicvf *nic, struct snd_queue *sq, int hdr_sqe, u8 subdesc_cnt) argument 572 nicvf_free_snd_queue(struct nicvf *nic, struct snd_queue *sq) argument 863 struct snd_queue *sq; local 1119 nicvf_get_sq_desc(struct snd_queue *sq, int desc_cnt) argument 1135 nicvf_rollback_sq_desc(struct snd_queue *sq, int qentry, int desc_cnt) argument 1143 nicvf_put_sq_desc(struct snd_queue *sq, int desc_cnt) argument 1153 nicvf_get_nxt_sqentry(struct snd_queue *sq, int qentry) argument 1160 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) argument 1180 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) argument 1206 nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num) argument 1222 nicvf_xdp_sq_add_hdr_subdesc(struct snd_queue *sq, int qentry, int subdesc_cnt, u64 data, int len) argument 1236 nicvf_xdp_sq_append_pkt(struct nicvf *nic, struct snd_queue *sq, u64 bufaddr, u64 dma_addr, u16 len) argument 1330 nicvf_sq_add_hdr_subdesc(struct nicvf *nic, struct snd_queue *sq, int qentry, int subdesc_cnt, struct sk_buff *skb, int len) argument 1419 nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, int size, u64 data) argument 1438 nicvf_sq_add_cqe_subdesc(struct snd_queue *sq, int qentry, int tso_sqe, struct sk_buff *skb) argument 1486 nicvf_sq_append_tso(struct nicvf *nic, struct snd_queue *sq, int sq_num, int qentry, struct sk_buff *skb) argument 1550 nicvf_sq_append_skb(struct nicvf *nic, struct snd_queue *sq, struct sk_buff *skb, u8 sq_num) argument 1827 struct snd_queue *sq; local [all...] |
/linux-master/drivers/soc/qcom/ |
H A D | qmi_interface.c | 19 struct sockaddr_qrtr *sq); 168 struct sockaddr_qrtr sq; local 178 sq.sq_family = qmi->sq.sq_family; 179 sq.sq_node = qmi->sq.sq_node; 180 sq.sq_port = QRTR_PORT_CTRL; 182 msg.msg_name = &sq; 183 msg.msg_namelen = sizeof(sq); 231 struct sockaddr_qrtr sq; local 390 qmi_invoke_handler(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *buf, size_t len) argument 442 struct sockaddr_qrtr sq; local 472 qmi_handle_message(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, const void *buf, size_t len) argument 527 struct sockaddr_qrtr sq; local 585 qmi_sock_create(struct qmi_handle *qmi, struct sockaddr_qrtr *sq) argument 739 qmi_send_message(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, int type, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument 792 qmi_send_request(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument 813 qmi_send_response(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument 833 qmi_send_indication(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument [all...] |
/linux-master/net/qrtr/ |
H A D | ns.c | 53 struct sockaddr_qrtr sq; member in struct:qrtr_lookup 195 static int announce_servers(struct sockaddr_qrtr *sq) argument 208 ret = service_announce_new(sq, srv); 288 lookup_notify(&lookup->sq, srv, false); 320 static int ctrl_cmd_hello(struct sockaddr_qrtr *sq) argument 324 ret = say_hello(sq); 328 return announce_servers(sq); 336 struct sockaddr_qrtr sq; local 364 sq.sq_family = AF_QIPCRTR; 365 sq 388 struct sockaddr_qrtr sq; local 589 struct sockaddr_qrtr sq; local 686 struct sockaddr_qrtr sq; local [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_qp.h | 57 #define HINIC_MIN_TX_NUM_WQEBBS(sq) \ 58 (HINIC_MIN_TX_WQE_SIZE((sq)->wq) / (sq)->wq->wqebb_size) 122 struct hinic_sq sq; member in struct:hinic_qp 133 struct hinic_sq *sq, u16 global_qid); 138 int hinic_init_sq(struct hinic_sq *sq, struct hinic_hwif *hwif, 142 void hinic_clean_sq(struct hinic_sq *sq); 149 int hinic_get_sq_free_wqebbs(struct hinic_sq *sq); 178 void hinic_sq_prepare_wqe(struct hinic_sq *sq, struct hinic_sq_wqe *wqe, 181 void hinic_sq_write_db(struct hinic_sq *sq, u1 [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_cmdq.c | 26 u64 db_data = FIELD_PREP(ERDMA_CMD_HDR_WQEBB_INDEX_MASK, cmdq->sq.pi); 28 *cmdq->sq.db_record = db_data; 91 struct erdma_cmdq_sq *sq = &cmdq->sq; local 94 sq->wqebb_cnt = SQEBB_COUNT(ERDMA_CMDQ_SQE_SIZE); 95 sq->depth = cmdq->max_outstandings * sq->wqebb_cnt; 97 buf_size = sq->depth << SQEBB_SHIFT; 99 sq->qbuf = 101 &sq [all...] |