Searched refs:sq (Results 1 - 25 of 239) sorted by relevance

12345678910

/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dtx.c46 static void mlx5e_xsk_tx_post_err(struct mlx5e_xdpsq *sq, argument
49 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc);
50 struct mlx5e_xdp_wqe_info *wi = &sq->db.wqe_info[pi];
56 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc);
57 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, *xdpi);
58 if (xp_tx_metadata_enabled(sq->xsk_pool))
59 mlx5e_xdpi_fifo_push(&sq
64 mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigned int budget) argument
[all...]
H A Dtx.h13 bool mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigned int budget);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_tx.c46 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) argument
52 mlx5e_dma_get(sq, --sq->dma_fifo_pc);
54 mlx5e_tx_dma_unmap(sq->pdev, last_pushed_dma);
119 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, argument
123 if (unlikely(mlx5e_ipsec_txwqe_build_eseg_csum(sq, skb, eseg)))
131 sq->stats->csum_partial_inner++;
134 sq->stats->csum_partial++;
139 sq->stats->csum_partial++;
142 sq
149 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb, int *hopbyhop) argument
177 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, unsigned char *skb_data, u16 headlen, struct mlx5_wqe_data_seg *dseg) argument
243 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel) argument
262 mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5e_tx_attr *attr) argument
342 mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) argument
350 mlx5e_tx_flush(struct mlx5e_txqsq *sq) argument
371 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, u8 num_dma, struct mlx5e_tx_wqe_info *wi, struct mlx5_wqe_ctrl_seg *cseg, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) argument
423 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) argument
511 mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) argument
519 mlx5e_tx_mpwqe_session_start(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) argument
543 mlx5e_tx_mpwqe_session_is_active(struct mlx5e_txqsq *sq) argument
548 mlx5e_tx_mpwqe_add_dseg(struct mlx5e_txqsq *sq, struct mlx5e_xmit_data *txd) argument
566 mlx5e_tx_mpwqe_session_complete(struct mlx5e_txqsq *sq) argument
598 mlx5e_sq_xmit_mpwqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) argument
647 mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq) argument
662 mlx5e_txwqe_build_eseg(struct mlx5e_priv *priv, struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5_wqe_eth_seg *eseg, u16 ihs) argument
679 struct mlx5e_txqsq *sq; local
732 mlx5e_tx_wi_dma_unmap(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) argument
744 mlx5e_consume_skb(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_cqe64 *cqe, int napi_budget) argument
762 mlx5e_tx_wi_consume_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, struct mlx5_cqe64 *cqe, int napi_budget) argument
774 mlx5e_txqsq_wake(struct mlx5e_txqsq *sq) argument
788 struct mlx5e_txqsq *sq; local
889 mlx5e_tx_wi_kfree_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi) argument
897 mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq) argument
972 mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_av *av, u32 dqpn, u32 dqkey, bool xmit_more) argument
[all...]
H A Den_txrx.c49 static void mlx5e_handle_tx_dim(struct mlx5e_txqsq *sq) argument
51 struct mlx5e_sq_stats *stats = sq->stats;
54 if (unlikely(!test_bit(MLX5E_SQ_STATE_DIM, &sq->state)))
57 dim_update_sample(sq->cq.event_ctr, stats->packets, stats->bytes, &dim_sample);
58 net_dim(&sq->dim, dim_sample);
73 void mlx5e_trigger_irq(struct mlx5e_icosq *sq) argument
75 struct mlx5_wq_cyc *wq = &sq->wq;
77 u16 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc);
79 sq->db.wqe_info[pi] = (struct mlx5e_icosq_wqe_info) {
84 nopwqe = mlx5e_post_nop(wq, sq
157 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); local
225 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); local
[all...]
/linux-master/drivers/nvme/target/
H A Dfabrics-cmd-auth.c17 struct nvmet_sq *sq = container_of(to_delayed_work(work), local
21 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid);
22 sq->dhchap_step = NVME_AUTH_DHCHAP_MESSAGE_NEGOTIATE;
23 sq->dhchap_tid = -1;
26 void nvmet_auth_sq_init(struct nvmet_sq *sq) argument
29 INIT_DELAYED_WORK(&sq->auth_expired_work, nvmet_auth_expired_work);
30 sq->authenticated = false;
31 sq
[all...]
H A Dauth.c222 void nvmet_auth_sq_free(struct nvmet_sq *sq) argument
224 cancel_delayed_work(&sq->auth_expired_work);
225 kfree(sq->dhchap_c1);
226 sq->dhchap_c1 = NULL;
227 kfree(sq->dhchap_c2);
228 sq->dhchap_c2 = NULL;
229 kfree(sq->dhchap_skey);
230 sq->dhchap_skey = NULL;
257 if (req->sq->ctrl->host_key &&
258 !req->sq
[all...]
/linux-master/tools/include/io_uring/
H A Dmini_liburing.h55 struct io_uring_sq sq; member in struct:io_uring
69 struct io_uring_sq *sq, struct io_uring_cq *cq)
75 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned int);
76 ptr = mmap(0, sq->ring_sz, PROT_READ | PROT_WRITE,
80 sq->khead = ptr + p->sq_off.head;
81 sq->ktail = ptr + p->sq_off.tail;
82 sq->kring_mask = ptr + p->sq_off.ring_mask;
83 sq->kring_entries = ptr + p->sq_off.ring_entries;
84 sq->kflags = ptr + p->sq_off.flags;
85 sq
68 io_uring_mmap(int fd, struct io_uring_params *p, struct io_uring_sq *sq, struct io_uring_cq *cq) argument
154 struct io_uring_sq *sq = &ring->sq; local
187 struct io_uring_sq *sq = &ring->sq; local
222 struct io_uring_sq *sq = &ring->sq; local
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dxdp.c61 mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq, argument
96 dma_addr = dma_map_single(sq->pdev, xdptxd->data, xdptxd->len,
98 if (dma_mapping_error(sq->pdev, dma_addr)) {
105 if (unlikely(!INDIRECT_CALL_2(sq->xmit_xdp_frame, mlx5e_xmit_xdp_frame_mpwqe,
106 mlx5e_xmit_xdp_frame, sq, xdptxd, 0, NULL)))
110 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo,
112 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo,
114 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo,
126 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd->len, DMA_BIDIRECTIONAL);
140 dma_sync_single_for_device(sq
349 mlx5e_xdpsq_get_next_pi(struct mlx5e_xdpsq *sq, u16 size) argument
378 mlx5e_xdp_mpwqe_session_start(struct mlx5e_xdpsq *sq) argument
400 mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq) argument
428 mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq) argument
450 mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result, struct xsk_tx_metadata *meta) argument
511 mlx5e_xmit_xdp_frame_check_stop_room(struct mlx5e_xdpsq *sq, int stop_room) argument
523 mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq) argument
529 mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result, struct xsk_tx_metadata *meta) argument
659 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, struct xdp_frame_bulk *bq, struct mlx5e_cq *cq, struct mlx5_cqe64 *cqe) argument
751 struct mlx5e_xdpsq *sq; local
819 mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq) argument
851 struct mlx5e_xdpsq *sq; local
968 mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw) argument
[all...]
H A Dreporter_tx.c22 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) argument
24 struct mlx5_core_dev *dev = sq->mdev;
30 if (sq->cc == sq->pc)
36 netdev_err(sq->netdev,
37 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n",
38 sq->sqn, sq->cc, sq
43 mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) argument
53 mlx5e_health_sq_put_sw_state(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq) argument
72 struct mlx5e_txqsq *sq; local
127 struct mlx5e_txqsq *sq; member in struct:mlx5e_tx_timeout_ctx
136 struct mlx5e_txqsq *sq; local
225 mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) argument
249 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) argument
341 struct mlx5e_txqsq *sq = &c->sq[tc]; local
366 struct mlx5e_txqsq *sq = ctx; local
434 struct mlx5e_txqsq *sq = &c->sq[tc]; local
442 struct mlx5e_txqsq *sq = &ptp_ch->ptpsq[tc].txqsq; local
470 mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq) argument
484 mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq) argument
[all...]
H A Dxdp.h108 void mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq);
110 void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq);
111 void mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw);
119 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq,
123 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq,
127 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq));
128 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq));
158 static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) argument
160 if (sq->doorbell_cseg) {
161 mlx5e_notify_hw(&sq
169 mlx5e_xdp_get_inline_state(struct mlx5e_xdpsq *sq, bool cur) argument
200 mlx5e_xdp_mpwqe_add_dseg(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdpsq_stats *stats) argument
[all...]
H A Dtxrx.h61 void mlx5e_trigger_irq(struct mlx5e_icosq *sq);
95 void mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq);
119 #define MLX5E_TX_FETCH_WQE(sq, pi) \
120 ((struct mlx5e_tx_wqe *)mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_tx_wqe)))
168 static inline u16 mlx5e_txqsq_get_next_pi(struct mlx5e_txqsq *sq, u16 size) argument
170 struct mlx5_wq_cyc *wq = &sq->wq;
173 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc);
178 wi = &sq->db.wqe_info[pi];
186 mlx5e_post_nop(wq, sq->sqn, &sq
230 mlx5e_icosq_get_next_pi(struct mlx5e_icosq *sq, u16 size) argument
285 mlx5e_dma_get(struct mlx5e_txqsq *sq, u32 i) argument
291 mlx5e_dma_push(struct mlx5e_txqsq *sq, dma_addr_t addr, u32 size, enum mlx5e_dma_map_type map_type) argument
496 mlx5e_icosq_can_post_wqe(struct mlx5e_icosq *sq, u16 wqe_size) argument
[all...]
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dqos_sq.c38 struct otx2_snd_queue *sq; local
76 sq = &qset->sq[qidx];
77 sq->sqb_count = 0;
78 sq->sqb_ptrs = kcalloc(num_sqbs, sizeof(*sq->sqb_ptrs), GFP_KERNEL);
79 if (!sq->sqb_ptrs) {
89 sq->sqb_ptrs[sq->sqb_count++] = (u64)bufptr;
96 if (!sq
120 struct otx2_snd_queue *sq; local
263 struct otx2_snd_queue *sq; local
[all...]
H A Dotx2_txrx.c116 struct otx2_snd_queue *sq,
124 sg = &sq->sg[snd_comp->sqe_id];
135 struct otx2_snd_queue *sq,
151 sg = &sq->sg[snd_comp->sqe_id];
157 timestamp = ((u64 *)sq->timestamps->base)[snd_comp->sqe_id];
451 struct otx2_snd_queue *sq; local
463 sq = &pfvf->qset.sq[qidx];
476 otx2_xdp_snd_pkt_handler(pfvf, sq, cqe);
478 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[qid
115 otx2_xdp_snd_pkt_handler(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_cqe_tx_s *cqe) argument
133 otx2_snd_pkt_handler(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, struct otx2_snd_queue *sq, struct nix_cqe_tx_s *cqe, int budget, int *tx_pkts, int *tx_bytes) argument
595 otx2_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx) argument
614 otx2_sqe_add_sg(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int num_segs, int *offset) argument
659 otx2_sqe_add_ext(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int *offset) argument
727 otx2_sqe_add_mem(struct otx2_snd_queue *sq, int *offset, int alg, u64 iova, int ptp_offset, u64 base_ns, bool udp_csum_crt) argument
750 otx2_sqe_add_hdr(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_sqe_hdr_s *sqe_hdr, struct sk_buff *skb, u16 qidx) argument
798 otx2_dma_map_tso_skb(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int sqe, int hdr_len) argument
831 otx2_tso_frag_dma_addr(struct otx2_snd_queue *sq, struct sk_buff *skb, int seg, u64 seg_addr, int hdr_len, int sqe) argument
849 otx2_sqe_tso_add_sg(struct otx2_snd_queue *sq, struct sg_list *list, int *offset) argument
880 otx2_sq_append_tso(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) argument
1078 otx2_set_txtstamp(struct otx2_nic *pfvf, struct sk_buff *skb, struct otx2_snd_queue *sq, int *offset) argument
1141 otx2_sq_append_skb(struct net_device *netdev, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) argument
1252 struct otx2_snd_queue *sq; local
1319 struct otx2_snd_queue *sq; local
1347 otx2_xdp_sqe_add_sg(struct otx2_snd_queue *sq, u64 dma_addr, int len, int *offset) argument
1370 struct otx2_snd_queue *sq; local
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dktls_utils.h62 #define MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi) \
64 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_static_params_wqe)))
66 #define MLX5E_TLS_FETCH_SET_PROGRESS_PARAMS_WQE(sq, pi) \
68 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_progress_params_wqe)))
70 #define MLX5E_TLS_FETCH_GET_PROGRESS_PARAMS_WQE(sq, pi) \
72 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_get_tls_progress_params_wqe)))
74 #define MLX5E_TLS_FETCH_DUMP_WQE(sq, pi) \
76 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_dump_wqe)))
H A Dktls_txrx.h19 bool mlx5e_ktls_handle_tx_skb(struct net_device *netdev, struct mlx5e_txqsq *sq,
27 struct mlx5e_icosq *sq);
29 void mlx5e_ktls_tx_handle_resync_dump_comp(struct mlx5e_txqsq *sq,
33 mlx5e_ktls_tx_try_handle_resync_dump_comp(struct mlx5e_txqsq *sq, argument
38 mlx5e_ktls_tx_handle_resync_dump_comp(sq, wi, dma_fifo_cc);
60 mlx5e_ktls_tx_try_handle_resync_dump_comp(struct mlx5e_txqsq *sq, argument
H A Dktls_rx.c129 static void icosq_fill_wi(struct mlx5e_icosq *sq, u16 pi, argument
132 sq->db.wqe_info[pi] = *wi;
136 post_static_params(struct mlx5e_icosq *sq, argument
144 if (unlikely(!mlx5e_icosq_can_post_wqe(sq, num_wqebbs)))
147 pi = mlx5e_icosq_get_next_pi(sq, num_wqebbs);
148 wqe = MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi);
149 mlx5e_ktls_build_static_params(wqe, sq->pc, sq->sqn, &priv_rx->crypto_info,
159 icosq_fill_wi(sq, pi, &wi);
160 sq
166 post_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx, u32 next_record_tcp_sn) argument
201 struct mlx5e_icosq *sq; local
252 resync_post_get_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx) argument
335 struct mlx5e_icosq *sq; local
367 struct mlx5e_icosq *sq; local
423 mlx5e_ktls_handle_get_psv_completion(struct mlx5e_icosq_wqe_info *wi, struct mlx5e_icosq *sq) argument
724 struct mlx5e_icosq *sq; local
[all...]
/linux-master/sound/oss/dmasound/
H A Ddmasound_core.c411 static int sq_allocate_buffers(struct sound_queue *sq, int num, int size) argument
415 if (sq->buffers)
417 sq->numBufs = num;
418 sq->bufSize = size;
419 sq->buffers = kmalloc_array (num, sizeof(char *), GFP_KERNEL);
420 if (!sq->buffers)
423 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL);
424 if (!sq->buffers[i]) {
426 dmasound.mach.dma_free(sq->buffers[i], size);
427 kfree(sq
435 sq_release_buffers(struct sound_queue *sq) argument
448 sq_setup(struct sound_queue *sq) argument
688 sq_init_waitqueue(struct sound_queue *sq) argument
707 sq_open2(struct sound_queue *sq, struct file *file, fmode_t mode, int numbufs, int bufsize) argument
953 set_queue_frags(struct sound_queue *sq, int bufs, int size) argument
[all...]
/linux-master/drivers/infiniband/hw/mlx5/
H A Dwr.h27 * @sq - SQ buffer.
33 static inline void *get_sq_edge(struct mlx5_ib_wq *sq, u32 idx) argument
38 (&sq->fbc,
39 mlx5_frag_buf_get_idx_last_contig_stride(&sq->fbc, idx));
46 * @sq: SQ buffer.
51 static inline void handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, argument
59 idx = (sq->cur_post + (wqe_sz >> 2)) & (sq->wqe_cnt - 1);
60 *cur_edge = get_sq_edge(sq, idx);
62 *seg = mlx5_frag_buf_get_wqe(&sq
74 mlx5r_memcpy_send_wqe(struct mlx5_ib_wq *sq, void **cur_edge, void **seg, u32 *wqe_sz, const void *src, size_t n) argument
[all...]
H A Dib_rep.h18 struct mlx5_ib_sq *sq,
27 struct mlx5_ib_sq *sq,
26 create_flow_rule_vport_sq(struct mlx5_ib_dev *dev, struct mlx5_ib_sq *sq, u32 port) argument
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_controlq.c8 (qinfo)->sq.head = prefix##_ATQH; \
9 (qinfo)->sq.tail = prefix##_ATQT; \
10 (qinfo)->sq.len = prefix##_ATQLEN; \
11 (qinfo)->sq.bah = prefix##_ATQBAH; \
12 (qinfo)->sq.bal = prefix##_ATQBAL; \
13 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \
14 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \
15 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \
16 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \
77 if (cq->sq
860 struct ice_ctl_q_ring *sq = &cq->sq; local
[all...]
/linux-master/drivers/net/ethernet/cavium/thunder/
H A Dnicvf_queues.c20 static inline void nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry,
505 struct snd_queue *sq, int q_len, int qidx)
509 err = nicvf_alloc_q_desc_mem(nic, &sq->dmem, q_len, SND_QUEUE_DESC_SIZE,
514 sq->desc = sq->dmem.base;
515 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL);
516 if (!sq->skbuff)
519 sq->head = 0;
520 sq->tail = 0;
521 sq
504 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) argument
554 nicvf_unmap_sndq_buffers(struct nicvf *nic, struct snd_queue *sq, int hdr_sqe, u8 subdesc_cnt) argument
572 nicvf_free_snd_queue(struct nicvf *nic, struct snd_queue *sq) argument
863 struct snd_queue *sq; local
1119 nicvf_get_sq_desc(struct snd_queue *sq, int desc_cnt) argument
1135 nicvf_rollback_sq_desc(struct snd_queue *sq, int qentry, int desc_cnt) argument
1143 nicvf_put_sq_desc(struct snd_queue *sq, int desc_cnt) argument
1153 nicvf_get_nxt_sqentry(struct snd_queue *sq, int qentry) argument
1160 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) argument
1180 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) argument
1206 nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num) argument
1222 nicvf_xdp_sq_add_hdr_subdesc(struct snd_queue *sq, int qentry, int subdesc_cnt, u64 data, int len) argument
1236 nicvf_xdp_sq_append_pkt(struct nicvf *nic, struct snd_queue *sq, u64 bufaddr, u64 dma_addr, u16 len) argument
1330 nicvf_sq_add_hdr_subdesc(struct nicvf *nic, struct snd_queue *sq, int qentry, int subdesc_cnt, struct sk_buff *skb, int len) argument
1419 nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, int size, u64 data) argument
1438 nicvf_sq_add_cqe_subdesc(struct snd_queue *sq, int qentry, int tso_sqe, struct sk_buff *skb) argument
1486 nicvf_sq_append_tso(struct nicvf *nic, struct snd_queue *sq, int sq_num, int qentry, struct sk_buff *skb) argument
1550 nicvf_sq_append_skb(struct nicvf *nic, struct snd_queue *sq, struct sk_buff *skb, u8 sq_num) argument
1827 struct snd_queue *sq; local
[all...]
/linux-master/drivers/soc/qcom/
H A Dqmi_interface.c19 struct sockaddr_qrtr *sq);
168 struct sockaddr_qrtr sq; local
178 sq.sq_family = qmi->sq.sq_family;
179 sq.sq_node = qmi->sq.sq_node;
180 sq.sq_port = QRTR_PORT_CTRL;
182 msg.msg_name = &sq;
183 msg.msg_namelen = sizeof(sq);
231 struct sockaddr_qrtr sq; local
390 qmi_invoke_handler(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, const void *buf, size_t len) argument
442 struct sockaddr_qrtr sq; local
472 qmi_handle_message(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, const void *buf, size_t len) argument
527 struct sockaddr_qrtr sq; local
585 qmi_sock_create(struct qmi_handle *qmi, struct sockaddr_qrtr *sq) argument
739 qmi_send_message(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, int type, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument
792 qmi_send_request(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument
813 qmi_send_response(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, struct qmi_txn *txn, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument
833 qmi_send_indication(struct qmi_handle *qmi, struct sockaddr_qrtr *sq, int msg_id, size_t len, const struct qmi_elem_info *ei, const void *c_struct) argument
[all...]
/linux-master/net/qrtr/
H A Dns.c53 struct sockaddr_qrtr sq; member in struct:qrtr_lookup
195 static int announce_servers(struct sockaddr_qrtr *sq) argument
208 ret = service_announce_new(sq, srv);
288 lookup_notify(&lookup->sq, srv, false);
320 static int ctrl_cmd_hello(struct sockaddr_qrtr *sq) argument
324 ret = say_hello(sq);
328 return announce_servers(sq);
336 struct sockaddr_qrtr sq; local
364 sq.sq_family = AF_QIPCRTR;
365 sq
388 struct sockaddr_qrtr sq; local
589 struct sockaddr_qrtr sq; local
686 struct sockaddr_qrtr sq; local
[all...]
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_qp.h57 #define HINIC_MIN_TX_NUM_WQEBBS(sq) \
58 (HINIC_MIN_TX_WQE_SIZE((sq)->wq) / (sq)->wq->wqebb_size)
122 struct hinic_sq sq; member in struct:hinic_qp
133 struct hinic_sq *sq, u16 global_qid);
138 int hinic_init_sq(struct hinic_sq *sq, struct hinic_hwif *hwif,
142 void hinic_clean_sq(struct hinic_sq *sq);
149 int hinic_get_sq_free_wqebbs(struct hinic_sq *sq);
178 void hinic_sq_prepare_wqe(struct hinic_sq *sq, struct hinic_sq_wqe *wqe,
181 void hinic_sq_write_db(struct hinic_sq *sq, u1
[all...]
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_cmdq.c26 u64 db_data = FIELD_PREP(ERDMA_CMD_HDR_WQEBB_INDEX_MASK, cmdq->sq.pi);
28 *cmdq->sq.db_record = db_data;
91 struct erdma_cmdq_sq *sq = &cmdq->sq; local
94 sq->wqebb_cnt = SQEBB_COUNT(ERDMA_CMDQ_SQE_SIZE);
95 sq->depth = cmdq->max_outstandings * sq->wqebb_cnt;
97 buf_size = sq->depth << SQEBB_SHIFT;
99 sq->qbuf =
101 &sq
[all...]

Completed in 404 milliseconds

12345678910