1/*-
2 * SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
3 *
4 * Copyright (c) 2015 - 2023 Intel Corporation
5 *
6 * This software is available to you under a choice of one of two
7 * licenses.  You may choose to be licensed under the terms of the GNU
8 * General Public License (GPL) Version 2, available from the file
9 * COPYING in the main directory of this source tree, or the
10 * OpenFabrics.org BSD license below:
11 *
12 *   Redistribution and use in source and binary forms, with or
13 *   without modification, are permitted provided that the following
14 *   conditions are met:
15 *
16 *    - Redistributions of source code must retain the above
17 *	copyright notice, this list of conditions and the following
18 *	disclaimer.
19 *
20 *    - Redistributions in binary form must reproduce the above
21 *	copyright notice, this list of conditions and the following
22 *	disclaimer in the documentation and/or other materials
23 *	provided with the distribution.
24 *
25 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32 * SOFTWARE.
33 */
34
35#ifndef IRDMA_TYPE_H
36#define IRDMA_TYPE_H
37
38#include "osdep.h"
39
40#include "irdma.h"
41#include "irdma_user.h"
42#include "irdma_hmc.h"
43#include "irdma_uda.h"
44#include "irdma_ws.h"
45#include "irdma_pble.h"
46
47enum irdma_debug_flag {
48	IRDMA_DEBUG_NONE	= 0x00000000,
49	IRDMA_DEBUG_ERR		= 0x00000001,
50	IRDMA_DEBUG_INIT	= 0x00000002,
51	IRDMA_DEBUG_DEV		= 0x00000004,
52	IRDMA_DEBUG_CM		= 0x00000008,
53	IRDMA_DEBUG_VERBS	= 0x00000010,
54	IRDMA_DEBUG_PUDA	= 0x00000020,
55	IRDMA_DEBUG_ILQ		= 0x00000040,
56	IRDMA_DEBUG_IEQ		= 0x00000080,
57	IRDMA_DEBUG_QP		= 0x00000100,
58	IRDMA_DEBUG_CQ		= 0x00000200,
59	IRDMA_DEBUG_MR		= 0x00000400,
60	IRDMA_DEBUG_PBLE	= 0x00000800,
61	IRDMA_DEBUG_WQE		= 0x00001000,
62	IRDMA_DEBUG_AEQ		= 0x00002000,
63	IRDMA_DEBUG_CQP		= 0x00004000,
64	IRDMA_DEBUG_HMC		= 0x00008000,
65	IRDMA_DEBUG_USER	= 0x00010000,
66	IRDMA_DEBUG_VIRT	= 0x00020000,
67	IRDMA_DEBUG_DCB		= 0x00040000,
68	IRDMA_DEBUG_CQE		= 0x00800000,
69	IRDMA_DEBUG_CLNT	= 0x01000000,
70	IRDMA_DEBUG_WS		= 0x02000000,
71	IRDMA_DEBUG_STATS	= 0x04000000,
72	IRDMA_DEBUG_ALL		= 0xFFFFFFFF,
73};
74
75#define RSVD_OFFSET 0xFFFFFFFF
76
77enum irdma_page_size {
78	IRDMA_PAGE_SIZE_4K = 0,
79	IRDMA_PAGE_SIZE_2M,
80	IRDMA_PAGE_SIZE_1G,
81};
82
83enum irdma_hdrct_flags {
84	DDP_LEN_FLAG  = 0x80,
85	DDP_HDR_FLAG  = 0x40,
86	RDMA_HDR_FLAG = 0x20,
87};
88
89enum irdma_term_layers {
90	LAYER_RDMA = 0,
91	LAYER_DDP  = 1,
92	LAYER_MPA  = 2,
93};
94
95enum irdma_term_error_types {
96	RDMAP_REMOTE_PROT = 1,
97	RDMAP_REMOTE_OP   = 2,
98	DDP_CATASTROPHIC  = 0,
99	DDP_TAGGED_BUF    = 1,
100	DDP_UNTAGGED_BUF  = 2,
101	DDP_LLP		  = 3,
102};
103
104enum irdma_term_rdma_errors {
105	RDMAP_INV_STAG		  = 0x00,
106	RDMAP_INV_BOUNDS	  = 0x01,
107	RDMAP_ACCESS		  = 0x02,
108	RDMAP_UNASSOC_STAG	  = 0x03,
109	RDMAP_TO_WRAP		  = 0x04,
110	RDMAP_INV_RDMAP_VER       = 0x05,
111	RDMAP_UNEXPECTED_OP       = 0x06,
112	RDMAP_CATASTROPHIC_LOCAL  = 0x07,
113	RDMAP_CATASTROPHIC_GLOBAL = 0x08,
114	RDMAP_CANT_INV_STAG       = 0x09,
115	RDMAP_UNSPECIFIED	  = 0xff,
116};
117
118enum irdma_term_ddp_errors {
119	DDP_CATASTROPHIC_LOCAL      = 0x00,
120	DDP_TAGGED_INV_STAG	    = 0x00,
121	DDP_TAGGED_BOUNDS	    = 0x01,
122	DDP_TAGGED_UNASSOC_STAG     = 0x02,
123	DDP_TAGGED_TO_WRAP	    = 0x03,
124	DDP_TAGGED_INV_DDP_VER      = 0x04,
125	DDP_UNTAGGED_INV_QN	    = 0x01,
126	DDP_UNTAGGED_INV_MSN_NO_BUF = 0x02,
127	DDP_UNTAGGED_INV_MSN_RANGE  = 0x03,
128	DDP_UNTAGGED_INV_MO	    = 0x04,
129	DDP_UNTAGGED_INV_TOO_LONG   = 0x05,
130	DDP_UNTAGGED_INV_DDP_VER    = 0x06,
131};
132
133enum irdma_term_mpa_errors {
134	MPA_CLOSED  = 0x01,
135	MPA_CRC     = 0x02,
136	MPA_MARKER  = 0x03,
137	MPA_REQ_RSP = 0x04,
138};
139
140enum irdma_hw_stats_index {
141	/* gen1 - 32-bit */
142	IRDMA_HW_STAT_INDEX_IP4RXDISCARD	= 0,
143	IRDMA_HW_STAT_INDEX_IP4RXTRUNC		= 1,
144	IRDMA_HW_STAT_INDEX_IP4TXNOROUTE	= 2,
145	IRDMA_HW_STAT_INDEX_IP6RXDISCARD	= 3,
146	IRDMA_HW_STAT_INDEX_IP6RXTRUNC		= 4,
147	IRDMA_HW_STAT_INDEX_IP6TXNOROUTE	= 5,
148	IRDMA_HW_STAT_INDEX_TCPRTXSEG		= 6,
149	IRDMA_HW_STAT_INDEX_TCPRXOPTERR		= 7,
150	IRDMA_HW_STAT_INDEX_TCPRXPROTOERR	= 8,
151	IRDMA_HW_STAT_INDEX_RXVLANERR		= 9,
152	/* gen1 - 64-bit */
153	IRDMA_HW_STAT_INDEX_IP4RXOCTS		= 10,
154	IRDMA_HW_STAT_INDEX_IP4RXPKTS		= 11,
155	IRDMA_HW_STAT_INDEX_IP4RXFRAGS		= 12,
156	IRDMA_HW_STAT_INDEX_IP4RXMCPKTS		= 13,
157	IRDMA_HW_STAT_INDEX_IP4TXOCTS		= 14,
158	IRDMA_HW_STAT_INDEX_IP4TXPKTS		= 15,
159	IRDMA_HW_STAT_INDEX_IP4TXFRAGS		= 16,
160	IRDMA_HW_STAT_INDEX_IP4TXMCPKTS		= 17,
161	IRDMA_HW_STAT_INDEX_IP6RXOCTS		= 18,
162	IRDMA_HW_STAT_INDEX_IP6RXPKTS		= 19,
163	IRDMA_HW_STAT_INDEX_IP6RXFRAGS		= 20,
164	IRDMA_HW_STAT_INDEX_IP6RXMCPKTS		= 21,
165	IRDMA_HW_STAT_INDEX_IP6TXOCTS		= 22,
166	IRDMA_HW_STAT_INDEX_IP6TXPKTS		= 23,
167	IRDMA_HW_STAT_INDEX_IP6TXFRAGS		= 24,
168	IRDMA_HW_STAT_INDEX_IP6TXMCPKTS		= 25,
169	IRDMA_HW_STAT_INDEX_TCPRXSEGS		= 26,
170	IRDMA_HW_STAT_INDEX_TCPTXSEG		= 27,
171	IRDMA_HW_STAT_INDEX_RDMARXRDS		= 28,
172	IRDMA_HW_STAT_INDEX_RDMARXSNDS		= 29,
173	IRDMA_HW_STAT_INDEX_RDMARXWRS		= 30,
174	IRDMA_HW_STAT_INDEX_RDMATXRDS		= 31,
175	IRDMA_HW_STAT_INDEX_RDMATXSNDS		= 32,
176	IRDMA_HW_STAT_INDEX_RDMATXWRS		= 33,
177	IRDMA_HW_STAT_INDEX_RDMAVBND		= 34,
178	IRDMA_HW_STAT_INDEX_RDMAVINV		= 35,
179	IRDMA_HW_STAT_INDEX_IP4RXMCOCTS		= 36,
180	IRDMA_HW_STAT_INDEX_IP4TXMCOCTS		= 37,
181	IRDMA_HW_STAT_INDEX_IP6RXMCOCTS		= 38,
182	IRDMA_HW_STAT_INDEX_IP6TXMCOCTS		= 39,
183	IRDMA_HW_STAT_INDEX_UDPRXPKTS		= 40,
184	IRDMA_HW_STAT_INDEX_UDPTXPKTS		= 41,
185	IRDMA_HW_STAT_INDEX_MAX_GEN_1		= 42, /* Must be same value as next entry */
186
187	/* gen2 - 64-bit */
188	IRDMA_HW_STAT_INDEX_RXNPECNMARKEDPKTS	= 42,
189
190	/* gen2 - 32-bit */
191	IRDMA_HW_STAT_INDEX_RXRPCNPHANDLED	= 43,
192	IRDMA_HW_STAT_INDEX_RXRPCNPIGNORED	= 44,
193	IRDMA_HW_STAT_INDEX_TXNPCNPSENT		= 45,
194	IRDMA_HW_STAT_INDEX_MAX_GEN_2		= 46,
195};
196
197#define IRDMA_MIN_FEATURES 2
198
199enum irdma_feature_type {
200	IRDMA_FEATURE_FW_INFO = 0,
201	IRDMA_HW_VERSION_INFO = 1,
202	IRDMA_QSETS_MAX       = 26,
203	IRDMA_MAX_FEATURES, /* Must be last entry */
204};
205
206enum irdma_sched_prio_type {
207	IRDMA_PRIO_WEIGHTED_RR     = 1,
208	IRDMA_PRIO_STRICT	   = 2,
209	IRDMA_PRIO_WEIGHTED_STRICT = 3,
210};
211
212enum irdma_vm_vf_type {
213	IRDMA_VF_TYPE = 0,
214	IRDMA_VM_TYPE,
215	IRDMA_PF_TYPE,
216};
217
218enum irdma_cqp_hmc_profile {
219	IRDMA_HMC_PROFILE_DEFAULT  = 1,
220	IRDMA_HMC_PROFILE_FAVOR_VF = 2,
221	IRDMA_HMC_PROFILE_EQUAL    = 3,
222};
223
224enum irdma_quad_entry_type {
225	IRDMA_QHASH_TYPE_TCP_ESTABLISHED = 1,
226	IRDMA_QHASH_TYPE_TCP_SYN,
227	IRDMA_QHASH_TYPE_UDP_UNICAST,
228	IRDMA_QHASH_TYPE_UDP_MCAST,
229	IRDMA_QHASH_TYPE_ROCE_MCAST,
230	IRDMA_QHASH_TYPE_ROCEV2_HW,
231};
232
233enum irdma_quad_hash_manage_type {
234	IRDMA_QHASH_MANAGE_TYPE_DELETE = 0,
235	IRDMA_QHASH_MANAGE_TYPE_ADD,
236	IRDMA_QHASH_MANAGE_TYPE_MODIFY,
237};
238
239enum irdma_syn_rst_handling {
240	IRDMA_SYN_RST_HANDLING_HW_TCP_SECURE = 0,
241	IRDMA_SYN_RST_HANDLING_HW_TCP,
242	IRDMA_SYN_RST_HANDLING_FW_TCP_SECURE,
243	IRDMA_SYN_RST_HANDLING_FW_TCP,
244};
245
246enum irdma_queue_type {
247	IRDMA_QUEUE_TYPE_SQ_RQ = 0,
248	IRDMA_QUEUE_TYPE_CQP,
249};
250
251struct irdma_sc_dev;
252struct irdma_vsi_pestat;
253
254struct irdma_dcqcn_cc_params {
255	u8 cc_cfg_valid;
256	u8 min_dec_factor;
257	u8 min_rate;
258	u8 dcqcn_f;
259	u16 rai_factor;
260	u16 hai_factor;
261	u16 dcqcn_t;
262	u32 dcqcn_b;
263	u32 rreduce_mperiod;
264};
265
266struct irdma_cqp_init_info {
267	u64 cqp_compl_ctx;
268	u64 host_ctx_pa;
269	u64 sq_pa;
270	struct irdma_sc_dev *dev;
271	struct irdma_cqp_quanta *sq;
272	struct irdma_dcqcn_cc_params dcqcn_params;
273	__le64 *host_ctx;
274	u64 *scratch_array;
275	u32 sq_size;
276	u16 hw_maj_ver;
277	u16 hw_min_ver;
278	u8 struct_ver;
279	u8 hmc_profile;
280	u8 ena_vf_count;
281	u8 ceqs_per_vf;
282	bool en_datacenter_tcp:1;
283	bool disable_packed:1;
284	bool rocev2_rto_policy:1;
285	bool en_rem_endpoint_trk:1;
286	enum irdma_protocol_used protocol_used;
287};
288
289struct irdma_terminate_hdr {
290	u8 layer_etype;
291	u8 error_code;
292	u8 hdrct;
293	u8 rsvd;
294};
295
296struct irdma_cqp_sq_wqe {
297	__le64 buf[IRDMA_CQP_WQE_SIZE];
298};
299
300struct irdma_sc_aeqe {
301	__le64 buf[IRDMA_AEQE_SIZE];
302};
303
304struct irdma_ceqe {
305	__le64 buf[IRDMA_CEQE_SIZE];
306};
307
308struct irdma_cqp_ctx {
309	__le64 buf[IRDMA_CQP_CTX_SIZE];
310};
311
312struct irdma_cq_shadow_area {
313	__le64 buf[IRDMA_SHADOW_AREA_SIZE];
314};
315
316struct irdma_dev_hw_stats {
317	u64 stats_val[IRDMA_GATHER_STATS_BUF_SIZE / sizeof(u64)];
318};
319
320struct irdma_gather_stats {
321	u64 val[IRDMA_GATHER_STATS_BUF_SIZE / sizeof(u64)];
322};
323
324struct irdma_hw_stat_map {
325	u16 byteoff;
326	u8 bitoff;
327	u64 bitmask;
328};
329
330struct irdma_stats_gather_info {
331	bool use_hmc_fcn_index:1;
332	bool use_stats_inst:1;
333	u16 hmc_fcn_index;
334	u16 stats_inst_index;
335	struct irdma_dma_mem stats_buff_mem;
336	void *gather_stats_va;
337	void *last_gather_stats_va;
338};
339
340struct irdma_vsi_pestat {
341	struct irdma_hw *hw;
342	struct irdma_dev_hw_stats hw_stats;
343	struct irdma_stats_gather_info gather_info;
344	struct OS_TIMER stats_timer;
345	struct irdma_sc_vsi *vsi;
346	spinlock_t lock; /* rdma stats lock */
347};
348
349struct irdma_hw {
350	u8 IOMEM *hw_addr;
351	u8 IOMEM *priv_hw_addr;
352	void *dev_context;
353	struct irdma_hmc_info hmc;
354};
355
356struct irdma_pfpdu {
357	struct list_head rxlist;
358	u32 rcv_nxt;
359	u32 fps;
360	u32 max_fpdu_data;
361	u32 nextseqnum;
362	u32 rcv_start_seq;
363	bool mode:1;
364	bool mpa_crc_err:1;
365	u8  marker_len;
366	u64 total_ieq_bufs;
367	u64 fpdu_processed;
368	u64 bad_seq_num;
369	u64 crc_err;
370	u64 no_tx_bufs;
371	u64 tx_err;
372	u64 out_of_order;
373	u64 pmode_count;
374	struct irdma_sc_ah *ah;
375	struct irdma_puda_buf *ah_buf;
376	spinlock_t lock; /* fpdu processing lock */
377	struct irdma_puda_buf *lastrcv_buf;
378};
379
380struct irdma_sc_pd {
381	struct irdma_sc_dev *dev;
382	u32 pd_id;
383	int abi_ver;
384};
385
386struct irdma_cqp_quanta {
387	__le64 elem[IRDMA_CQP_WQE_SIZE];
388};
389
390struct irdma_sc_cqp {
391	u32 size;
392	u64 sq_pa;
393	u64 host_ctx_pa;
394	void *back_cqp;
395	struct irdma_sc_dev *dev;
396	int (*process_cqp_sds)(struct irdma_sc_dev *dev,
397			       struct irdma_update_sds_info *info);
398	struct irdma_dma_mem sdbuf;
399	struct irdma_ring sq_ring;
400	struct irdma_cqp_quanta *sq_base;
401	struct irdma_dcqcn_cc_params dcqcn_params;
402	__le64 *host_ctx;
403	u64 *scratch_array;
404	u64 requested_ops;
405	atomic64_t completed_ops;
406	u32 cqp_id;
407	u32 sq_size;
408	u32 hw_sq_size;
409	u16 hw_maj_ver;
410	u16 hw_min_ver;
411	u8 struct_ver;
412	u8 polarity;
413	u8 hmc_profile;
414	u8 ena_vf_count;
415	u8 timeout_count;
416	u8 ceqs_per_vf;
417	bool en_datacenter_tcp:1;
418	bool disable_packed:1;
419	bool rocev2_rto_policy:1;
420	bool en_rem_endpoint_trk:1;
421	enum irdma_protocol_used protocol_used;
422};
423
424struct irdma_sc_aeq {
425	u32 size;
426	u64 aeq_elem_pa;
427	struct irdma_sc_dev *dev;
428	struct irdma_sc_aeqe *aeqe_base;
429	void *pbl_list;
430	u32 elem_cnt;
431	struct irdma_ring aeq_ring;
432	u8 pbl_chunk_size;
433	u32 first_pm_pbl_idx;
434	u32 msix_idx;
435	u8 polarity;
436	bool virtual_map:1;
437};
438
439struct irdma_sc_ceq {
440	u32 size;
441	u64 ceq_elem_pa;
442	struct irdma_sc_dev *dev;
443	struct irdma_ceqe *ceqe_base;
444	void *pbl_list;
445	u32 ceq_id;
446	u32 elem_cnt;
447	struct irdma_ring ceq_ring;
448	u8 pbl_chunk_size;
449	u8 tph_val;
450	u32 first_pm_pbl_idx;
451	u8 polarity;
452	struct irdma_sc_vsi *vsi;
453	struct irdma_sc_cq **reg_cq;
454	u32 reg_cq_size;
455	spinlock_t req_cq_lock; /* protect access to reg_cq array */
456	bool virtual_map:1;
457	bool tph_en:1;
458	bool itr_no_expire:1;
459};
460
461struct irdma_sc_cq {
462	struct irdma_cq_uk cq_uk;
463	u64 cq_pa;
464	u64 shadow_area_pa;
465	struct irdma_sc_dev *dev;
466	struct irdma_sc_vsi *vsi;
467	void *pbl_list;
468	void *back_cq;
469	u32 ceq_id;
470	u32 shadow_read_threshold;
471	u8 pbl_chunk_size;
472	u8 cq_type;
473	u8 tph_val;
474	u32 first_pm_pbl_idx;
475	bool ceqe_mask:1;
476	bool virtual_map:1;
477	bool check_overflow:1;
478	bool ceq_id_valid:1;
479	bool tph_en:1;
480};
481
482struct irdma_sc_qp {
483	struct irdma_qp_uk qp_uk;
484	u64 sq_pa;
485	u64 rq_pa;
486	u64 hw_host_ctx_pa;
487	u64 shadow_area_pa;
488	u64 q2_pa;
489	struct irdma_sc_dev *dev;
490	struct irdma_sc_vsi *vsi;
491	struct irdma_sc_pd *pd;
492	__le64 *hw_host_ctx;
493	void *llp_stream_handle;
494	struct irdma_pfpdu pfpdu;
495	u32 ieq_qp;
496	u8 *q2_buf;
497	u64 qp_compl_ctx;
498	u32 push_idx;
499	u16 qs_handle;
500	u16 push_offset;
501	u8 flush_wqes_count;
502	u8 sq_tph_val;
503	u8 rq_tph_val;
504	u8 qp_state;
505	u8 hw_sq_size;
506	u8 hw_rq_size;
507	u8 src_mac_addr_idx;
508	bool on_qoslist:1;
509	bool ieq_pass_thru:1;
510	bool sq_tph_en:1;
511	bool rq_tph_en:1;
512	bool rcv_tph_en:1;
513	bool xmit_tph_en:1;
514	bool virtual_map:1;
515	bool flush_sq:1;
516	bool flush_rq:1;
517	bool sq_flush_code:1;
518	bool rq_flush_code:1;
519	enum irdma_flush_opcode flush_code;
520	enum irdma_qp_event_type event_type;
521	u8 term_flags;
522	u8 user_pri;
523	struct list_head list;
524};
525
526struct irdma_stats_inst_info {
527	u16 hmc_fn_id;
528	u16 stats_idx;
529	bool use_hmc_fcn_index:1;
530};
531
532struct irdma_up_info {
533	u8 map[8];
534	u8 cnp_up_override;
535	u16 hmc_fcn_idx;
536	bool use_vlan:1;
537	bool use_cnp_up_override:1;
538};
539
540#define IRDMA_MAX_WS_NODES      0x3FF
541#define IRDMA_WS_NODE_INVALID	0xFFFF
542
543struct irdma_ws_node_info {
544	u16 id;
545	u16 vsi;
546	u16 parent_id;
547	u16 qs_handle;
548	bool type_leaf:1;
549	bool enable:1;
550	u8 prio_type;
551	u8 tc;
552	u8 weight;
553};
554
555struct irdma_hmc_fpm_misc {
556	u32 max_ceqs;
557	u32 max_sds;
558	u32 xf_block_size;
559	u32 q1_block_size;
560	u32 ht_multiplier;
561	u32 timer_bucket;
562	u32 rrf_block_size;
563	u32 ooiscf_block_size;
564};
565
566#define IRDMA_LEAF_DEFAULT_REL_BW		64
567#define IRDMA_PARENT_DEFAULT_REL_BW		1
568
569struct irdma_qos {
570	struct list_head qplist;
571	struct mutex qos_mutex; /* protect QoS attributes per QoS level */
572	u32 l2_sched_node_id;
573	u16 qs_handle;
574	u8 traffic_class;
575	u8 rel_bw;
576	u8 prio_type;
577	bool valid:1;
578};
579
580struct irdma_config_check {
581	bool config_ok:1;
582	bool lfc_set:1;
583	bool pfc_set:1;
584	u8 traffic_class;
585	u16 qs_handle;
586};
587
588#define IRDMA_INVALID_STATS_IDX 0xff
589struct irdma_sc_vsi {
590	u16 vsi_idx;
591	struct irdma_sc_dev *dev;
592	void *back_vsi;
593	u32 ilq_count;
594	struct irdma_virt_mem ilq_mem;
595	struct irdma_puda_rsrc *ilq;
596	u32 ieq_count;
597	struct irdma_virt_mem ieq_mem;
598	struct irdma_puda_rsrc *ieq;
599	u32 exception_lan_q;
600	u16 mtu;
601	enum irdma_vm_vf_type vm_vf_type;
602	bool stats_inst_alloc:1;
603	bool tc_change_pending:1;
604	bool mtu_change_pending:1;
605	struct irdma_vsi_pestat *pestat;
606	atomic_t qp_suspend_reqs;
607	int (*register_qset)(struct irdma_sc_vsi *vsi,
608			     struct irdma_ws_node *tc_node);
609	void (*unregister_qset)(struct irdma_sc_vsi *vsi,
610				struct irdma_ws_node *tc_node);
611	struct irdma_config_check cfg_check[IRDMA_MAX_USER_PRIORITY];
612	bool tc_print_warning[IRDMA_MAX_TRAFFIC_CLASS];
613	u8 qos_rel_bw;
614	u8 qos_prio_type;
615	u16 stats_idx;
616	u8 dscp_map[IRDMA_DSCP_NUM_VAL];
617	struct irdma_qos qos[IRDMA_MAX_USER_PRIORITY];
618	bool dscp_mode:1;
619};
620
621struct irdma_sc_dev {
622	struct list_head cqp_cmd_head; /* head of the CQP command list */
623	spinlock_t cqp_lock; /* protect CQP list access */
624	struct irdma_dma_mem vf_fpm_query_buf[IRDMA_MAX_PE_ENA_VF_COUNT];
625	u64 fpm_query_buf_pa;
626	u64 fpm_commit_buf_pa;
627	__le64 *fpm_query_buf;
628	__le64 *fpm_commit_buf;
629	struct irdma_hw *hw;
630	u32 IOMEM *wqe_alloc_db;
631	u32 IOMEM *cq_arm_db;
632	u32 IOMEM *aeq_alloc_db;
633	u32 IOMEM *cqp_db;
634	u32 IOMEM *cq_ack_db;
635	u32 IOMEM *hw_regs[IRDMA_MAX_REGS];
636	u32 ceq_itr;   /* Interrupt throttle, usecs between interrupts: 0 disabled. 2 - 8160 */
637	u64 hw_masks[IRDMA_MAX_MASKS];
638	u8 hw_shifts[IRDMA_MAX_SHIFTS];
639	const struct irdma_hw_stat_map *hw_stats_map;
640	u64 feature_info[IRDMA_MAX_FEATURES];
641	u64 cqp_cmd_stats[IRDMA_MAX_CQP_OPS];
642	struct irdma_hw_attrs hw_attrs;
643	struct irdma_hmc_info *hmc_info;
644	struct irdma_sc_cqp *cqp;
645	struct irdma_sc_aeq *aeq;
646	struct irdma_sc_ceq *ceq[IRDMA_CEQ_MAX_COUNT];
647	struct irdma_sc_cq *ccq;
648	const struct irdma_irq_ops *irq_ops;
649	struct irdma_hmc_fpm_misc hmc_fpm_misc;
650	struct irdma_ws_node *ws_tree_root;
651	struct mutex ws_mutex; /* ws tree mutex */
652	u32 debug_mask;
653	u16 num_vfs;
654	u16 hmc_fn_id;
655	bool ceq_valid:1;
656	u8 pci_rev;
657	int (*ws_add)(struct irdma_sc_vsi *vsi, u8 user_pri);
658	void (*ws_remove)(struct irdma_sc_vsi *vsi, u8 user_pri);
659	void (*ws_reset)(struct irdma_sc_vsi *vsi);
660};
661
662struct irdma_modify_cq_info {
663	u64 cq_pa;
664	struct irdma_cqe *cq_base;
665	u32 cq_size;
666	u32 shadow_read_threshold;
667	u8 pbl_chunk_size;
668	u32 first_pm_pbl_idx;
669	bool virtual_map:1;
670	bool check_overflow:1;
671	bool cq_resize:1;
672};
673
674struct irdma_create_qp_info {
675	bool ord_valid:1;
676	bool tcp_ctx_valid:1;
677	bool cq_num_valid:1;
678	bool arp_cache_idx_valid:1;
679	bool mac_valid:1;
680	bool force_lpb:1;
681	u8 next_iwarp_state;
682};
683
684struct irdma_modify_qp_info {
685	u64 rx_win0;
686	u64 rx_win1;
687	u16 new_mss;
688	u8 next_iwarp_state;
689	u8 curr_iwarp_state;
690	u8 termlen;
691	bool ord_valid:1;
692	bool tcp_ctx_valid:1;
693	bool udp_ctx_valid:1;
694	bool cq_num_valid:1;
695	bool arp_cache_idx_valid:1;
696	bool reset_tcp_conn:1;
697	bool remove_hash_idx:1;
698	bool dont_send_term:1;
699	bool dont_send_fin:1;
700	bool cached_var_valid:1;
701	bool mss_change:1;
702	bool force_lpb:1;
703	bool mac_valid:1;
704};
705
706struct irdma_ccq_cqe_info {
707	struct irdma_sc_cqp *cqp;
708	u64 scratch;
709	u32 op_ret_val;
710	u16 maj_err_code;
711	u16 min_err_code;
712	u8 op_code;
713	bool error:1;
714};
715
716struct irdma_qos_tc_info {
717	u64 tc_ctx;
718	u8 rel_bw;
719	u8 prio_type;
720	u8 egress_virt_up;
721	u8 ingress_virt_up;
722};
723
724struct irdma_l2params {
725	struct irdma_qos_tc_info tc_info[IRDMA_MAX_USER_PRIORITY];
726	u32 num_apps;
727	u16 qs_handle_list[IRDMA_MAX_USER_PRIORITY];
728	u16 mtu;
729	u8 up2tc[IRDMA_MAX_USER_PRIORITY];
730	u8 dscp_map[IRDMA_DSCP_NUM_VAL];
731	u8 num_tc;
732	u8 vsi_rel_bw;
733	u8 vsi_prio_type;
734	bool mtu_changed:1;
735	bool tc_changed:1;
736	bool dscp_mode:1;
737};
738
739struct irdma_vsi_init_info {
740	struct irdma_sc_dev *dev;
741	void *back_vsi;
742	struct irdma_l2params *params;
743	u16 exception_lan_q;
744	u16 pf_data_vsi_num;
745	enum irdma_vm_vf_type vm_vf_type;
746	int (*register_qset)(struct irdma_sc_vsi *vsi,
747			     struct irdma_ws_node *tc_node);
748	void (*unregister_qset)(struct irdma_sc_vsi *vsi,
749				struct irdma_ws_node *tc_node);
750};
751
752struct irdma_vsi_stats_info {
753	struct irdma_vsi_pestat *pestat;
754	u8 fcn_id;
755	bool alloc_stats_inst:1;
756};
757
758struct irdma_device_init_info {
759	u64 fpm_query_buf_pa;
760	u64 fpm_commit_buf_pa;
761	__le64 *fpm_query_buf;
762	__le64 *fpm_commit_buf;
763	struct irdma_hw *hw;
764	void IOMEM *bar0;
765	u16 max_vfs;
766	u16 hmc_fn_id;
767	u32 debug_mask;
768};
769
770struct irdma_ceq_init_info {
771	u64 ceqe_pa;
772	struct irdma_sc_dev *dev;
773	u64 *ceqe_base;
774	void *pbl_list;
775	u32 elem_cnt;
776	u32 ceq_id;
777	bool virtual_map:1;
778	bool tph_en:1;
779	bool itr_no_expire:1;
780	u8 pbl_chunk_size;
781	u8 tph_val;
782	u32 first_pm_pbl_idx;
783	struct irdma_sc_vsi *vsi;
784	struct irdma_sc_cq **reg_cq;
785};
786
787struct irdma_aeq_init_info {
788	u64 aeq_elem_pa;
789	struct irdma_sc_dev *dev;
790	u32 *aeqe_base;
791	void *pbl_list;
792	u32 elem_cnt;
793	bool virtual_map:1;
794	u8 pbl_chunk_size;
795	u32 first_pm_pbl_idx;
796	u32 msix_idx;
797};
798
799struct irdma_ccq_init_info {
800	u64 cq_pa;
801	u64 shadow_area_pa;
802	struct irdma_sc_dev *dev;
803	struct irdma_cqe *cq_base;
804	__le64 *shadow_area;
805	void *pbl_list;
806	u32 num_elem;
807	u32 ceq_id;
808	u32 shadow_read_threshold;
809	bool ceqe_mask:1;
810	bool ceq_id_valid:1;
811	bool avoid_mem_cflct:1;
812	bool virtual_map:1;
813	bool tph_en:1;
814	u8 tph_val;
815	u8 pbl_chunk_size;
816	u32 first_pm_pbl_idx;
817	struct irdma_sc_vsi *vsi;
818};
819
820struct irdma_udp_offload_info {
821	bool ipv4:1;
822	bool insert_vlan_tag:1;
823	u8 ttl;
824	u8 tos;
825	u16 src_port;
826	u16 dst_port;
827	u32 dest_ip_addr[4];
828	u32 snd_mss;
829	u16 vlan_tag;
830	u16 arp_idx;
831	u32 flow_label;
832	u8 udp_state;
833	u32 psn_nxt;
834	u32 lsn;
835	u32 epsn;
836	u32 psn_max;
837	u32 psn_una;
838	u32 local_ipaddr[4];
839	u32 cwnd;
840	u8 rexmit_thresh;
841	u8 rnr_nak_thresh;
842};
843
844struct irdma_roce_offload_info {
845	u16 p_key;
846	u32 err_rq_idx;
847	u32 qkey;
848	u32 dest_qp;
849	u8 roce_tver;
850	u8 ack_credits;
851	u8 err_rq_idx_valid;
852	u32 pd_id;
853	u16 ord_size;
854	u16 ird_size;
855	bool is_qp1:1;
856	bool udprivcq_en:1;
857	bool dcqcn_en:1;
858	bool rcv_no_icrc:1;
859	bool wr_rdresp_en:1;
860	bool fast_reg_en:1;
861	bool priv_mode_en:1;
862	bool rd_en:1;
863	bool timely_en:1;
864	bool dctcp_en:1;
865	bool fw_cc_enable:1;
866	bool use_stats_inst:1;
867	u16 t_high;
868	u16 t_low;
869	u8 last_byte_sent;
870	u8 mac_addr[ETHER_ADDR_LEN];
871	u8 rtomin;
872};
873
874struct irdma_iwarp_offload_info {
875	u16 rcv_mark_offset;
876	u16 snd_mark_offset;
877	u8 ddp_ver;
878	u8 rdmap_ver;
879	u8 iwarp_mode;
880	u32 err_rq_idx;
881	u32 pd_id;
882	u16 ord_size;
883	u16 ird_size;
884	bool ib_rd_en:1;
885	bool align_hdrs:1;
886	bool rcv_no_mpa_crc:1;
887	bool err_rq_idx_valid:1;
888	bool snd_mark_en:1;
889	bool rcv_mark_en:1;
890	bool wr_rdresp_en:1;
891	bool fast_reg_en:1;
892	bool priv_mode_en:1;
893	bool rd_en:1;
894	bool timely_en:1;
895	bool use_stats_inst:1;
896	bool ecn_en:1;
897	bool dctcp_en:1;
898	u16 t_high;
899	u16 t_low;
900	u8 last_byte_sent;
901	u8 mac_addr[ETHER_ADDR_LEN];
902	u8 rtomin;
903};
904
905struct irdma_tcp_offload_info {
906	bool ipv4:1;
907	bool no_nagle:1;
908	bool insert_vlan_tag:1;
909	bool time_stamp:1;
910	bool drop_ooo_seg:1;
911	bool avoid_stretch_ack:1;
912	bool wscale:1;
913	bool ignore_tcp_opt:1;
914	bool ignore_tcp_uns_opt:1;
915	u8 cwnd_inc_limit;
916	u8 dup_ack_thresh;
917	u8 ttl;
918	u8 src_mac_addr_idx;
919	u8 tos;
920	u16 src_port;
921	u16 dst_port;
922	u32 dest_ip_addr[4];
923	//u32 dest_ip_addr0;
924	//u32 dest_ip_addr1;
925	//u32 dest_ip_addr2;
926	//u32 dest_ip_addr3;
927	u32 snd_mss;
928	u16 syn_rst_handling;
929	u16 vlan_tag;
930	u16 arp_idx;
931	u32 flow_label;
932	u8 tcp_state;
933	u8 snd_wscale;
934	u8 rcv_wscale;
935	u32 time_stamp_recent;
936	u32 time_stamp_age;
937	u32 snd_nxt;
938	u32 snd_wnd;
939	u32 rcv_nxt;
940	u32 rcv_wnd;
941	u32 snd_max;
942	u32 snd_una;
943	u32 srtt;
944	u32 rtt_var;
945	u32 ss_thresh;
946	u32 cwnd;
947	u32 snd_wl1;
948	u32 snd_wl2;
949	u32 max_snd_window;
950	u8 rexmit_thresh;
951	u32 local_ipaddr[4];
952};
953
954struct irdma_qp_host_ctx_info {
955	u64 qp_compl_ctx;
956	union {
957		struct irdma_tcp_offload_info *tcp_info;
958		struct irdma_udp_offload_info *udp_info;
959	};
960	union {
961		struct irdma_iwarp_offload_info *iwarp_info;
962		struct irdma_roce_offload_info *roce_info;
963	};
964	u32 send_cq_num;
965	u32 rcv_cq_num;
966	u32 rem_endpoint_idx;
967	u16 stats_idx;
968	bool tcp_info_valid:1;
969	bool iwarp_info_valid:1;
970	bool stats_idx_valid:1;
971	u8 user_pri;
972};
973
974struct irdma_aeqe_info {
975	u64 compl_ctx;
976	u32 qp_cq_id;
977	u32 wqe_idx;
978	u16 ae_id;
979	u8 tcp_state;
980	u8 iwarp_state;
981	bool qp:1;
982	bool cq:1;
983	bool sq:1;
984	bool rq:1;
985	bool in_rdrsp_wr:1;
986	bool out_rdrsp:1;
987	bool aeqe_overflow:1;
988	/* This flag is used to determine if we should pass the rq tail
989	 * in the QP context for FW/HW. It is set when ae_src is rq for GEN1/GEN2
990	 * And additionally set for inbound atomic, read and write for GEN3
991	 */
992	bool err_rq_idx_valid:1;
993	u8 q2_data_written;
994	u8 ae_src;
995};
996
997struct irdma_allocate_stag_info {
998	u64 total_len;
999	u64 first_pm_pbl_idx;
1000	u32 chunk_size;
1001	u32 stag_idx;
1002	u32 page_size;
1003	u32 pd_id;
1004	u16 access_rights;
1005	bool remote_access:1;
1006	bool use_hmc_fcn_index:1;
1007	bool use_pf_rid:1;
1008	bool all_memory:1;
1009	u16 hmc_fcn_index;
1010};
1011
1012struct irdma_mw_alloc_info {
1013	u32 mw_stag_index;
1014	u32 page_size;
1015	u32 pd_id;
1016	bool remote_access:1;
1017	bool mw_wide:1;
1018	bool mw1_bind_dont_vldt_key:1;
1019};
1020
1021struct irdma_reg_ns_stag_info {
1022	u64 reg_addr_pa;
1023	u64 va;
1024	u64 total_len;
1025	u32 page_size;
1026	u32 chunk_size;
1027	u32 first_pm_pbl_index;
1028	enum irdma_addressing_type addr_type;
1029	irdma_stag_index stag_idx;
1030	u16 access_rights;
1031	u32 pd_id;
1032	irdma_stag_key stag_key;
1033	bool use_hmc_fcn_index:1;
1034	u16 hmc_fcn_index;
1035	bool use_pf_rid:1;
1036	bool all_memory:1;
1037};
1038
1039struct irdma_fast_reg_stag_info {
1040	u64 wr_id;
1041	u64 reg_addr_pa;
1042	u64 fbo;
1043	void *va;
1044	u64 total_len;
1045	u32 page_size;
1046	u32 chunk_size;
1047	u32 first_pm_pbl_index;
1048	enum irdma_addressing_type addr_type;
1049	irdma_stag_index stag_idx;
1050	u16 access_rights;
1051	u32 pd_id;
1052	irdma_stag_key stag_key;
1053	bool local_fence:1;
1054	bool read_fence:1;
1055	bool signaled:1;
1056	bool push_wqe:1;
1057	bool use_hmc_fcn_index:1;
1058	u16 hmc_fcn_index;
1059	bool use_pf_rid:1;
1060	bool defer_flag:1;
1061};
1062
1063struct irdma_dealloc_stag_info {
1064	u32 stag_idx;
1065	u32 pd_id;
1066	bool mr:1;
1067	bool dealloc_pbl:1;
1068};
1069
1070struct irdma_register_shared_stag {
1071	u64 va;
1072	enum irdma_addressing_type addr_type;
1073	irdma_stag_index new_stag_idx;
1074	irdma_stag_index parent_stag_idx;
1075	u32 access_rights;
1076	u32 pd_id;
1077	u32 page_size;
1078	irdma_stag_key new_stag_key;
1079};
1080
1081struct irdma_qp_init_info {
1082	struct irdma_qp_uk_init_info qp_uk_init_info;
1083	struct irdma_sc_pd *pd;
1084	struct irdma_sc_vsi *vsi;
1085	__le64 *host_ctx;
1086	u8 *q2;
1087	u64 sq_pa;
1088	u64 rq_pa;
1089	u64 host_ctx_pa;
1090	u64 q2_pa;
1091	u64 shadow_area_pa;
1092	u8 sq_tph_val;
1093	u8 rq_tph_val;
1094	bool sq_tph_en:1;
1095	bool rq_tph_en:1;
1096	bool rcv_tph_en:1;
1097	bool xmit_tph_en:1;
1098	bool virtual_map:1;
1099};
1100
1101struct irdma_cq_init_info {
1102	struct irdma_sc_dev *dev;
1103	u64 cq_base_pa;
1104	u64 shadow_area_pa;
1105	u32 ceq_id;
1106	u32 shadow_read_threshold;
1107	u8 pbl_chunk_size;
1108	u32 first_pm_pbl_idx;
1109	bool virtual_map:1;
1110	bool ceqe_mask:1;
1111	bool ceq_id_valid:1;
1112	bool tph_en:1;
1113	u8 tph_val;
1114	u8 type;
1115	struct irdma_cq_uk_init_info cq_uk_init_info;
1116	struct irdma_sc_vsi *vsi;
1117};
1118
1119struct irdma_upload_context_info {
1120	u64 buf_pa;
1121	u32 qp_id;
1122	u8 qp_type;
1123	bool freeze_qp:1;
1124	bool raw_format:1;
1125};
1126
1127struct irdma_local_mac_entry_info {
1128	u8 mac_addr[6];
1129	u16 entry_idx;
1130};
1131
1132struct irdma_add_arp_cache_entry_info {
1133	u8 mac_addr[ETHER_ADDR_LEN];
1134	u32 reach_max;
1135	u16 arp_index;
1136	bool permanent:1;
1137};
1138
1139struct irdma_apbvt_info {
1140	u16 port;
1141	bool add:1;
1142};
1143
1144struct irdma_qhash_table_info {
1145	struct irdma_sc_vsi *vsi;
1146	enum irdma_quad_hash_manage_type manage;
1147	enum irdma_quad_entry_type entry_type;
1148	bool vlan_valid:1;
1149	bool ipv4_valid:1;
1150	u8 mac_addr[ETHER_ADDR_LEN];
1151	u16 vlan_id;
1152	u8 user_pri;
1153	u32 qp_num;
1154	u32 dest_ip[4];
1155	u32 src_ip[4];
1156	u16 dest_port;
1157	u16 src_port;
1158};
1159
1160struct irdma_cqp_manage_push_page_info {
1161	u32 push_idx;
1162	u16 qs_handle;
1163	u8 free_page;
1164	u8 push_page_type;
1165};
1166
1167struct irdma_qp_flush_info {
1168	u16 sq_minor_code;
1169	u16 sq_major_code;
1170	u16 rq_minor_code;
1171	u16 rq_major_code;
1172	u16 ae_code;
1173	u8 ae_src;
1174	bool sq:1;
1175	bool rq:1;
1176	bool userflushcode:1;
1177	bool generate_ae:1;
1178};
1179
1180struct irdma_gen_ae_info {
1181	u16 ae_code;
1182	u8 ae_src;
1183};
1184
1185struct irdma_cqp_timeout {
1186	u64 compl_cqp_cmds;
1187	u32 count;
1188};
1189
1190struct irdma_irq_ops {
1191	void (*irdma_cfg_aeq)(struct irdma_sc_dev *dev, u32 idx, bool enable);
1192	void (*irdma_cfg_ceq)(struct irdma_sc_dev *dev, u32 ceq_id, u32 idx,
1193			      bool enable);
1194	void (*irdma_dis_irq)(struct irdma_sc_dev *dev, u32 idx);
1195	void (*irdma_en_irq)(struct irdma_sc_dev *dev, u32 idx);
1196};
1197
1198void irdma_sc_ccq_arm(struct irdma_sc_cq *ccq);
1199int irdma_sc_ccq_create(struct irdma_sc_cq *ccq, u64 scratch,
1200			bool check_overflow, bool post_sq);
1201int irdma_sc_ccq_destroy(struct irdma_sc_cq *ccq, u64 scratch, bool post_sq);
1202int irdma_sc_ccq_get_cqe_info(struct irdma_sc_cq *ccq,
1203			      struct irdma_ccq_cqe_info *info);
1204int irdma_sc_ccq_init(struct irdma_sc_cq *ccq,
1205		      struct irdma_ccq_init_info *info);
1206
1207int irdma_sc_cceq_create(struct irdma_sc_ceq *ceq, u64 scratch);
1208int irdma_sc_cceq_destroy_done(struct irdma_sc_ceq *ceq);
1209
1210int irdma_sc_ceq_destroy(struct irdma_sc_ceq *ceq, u64 scratch, bool post_sq);
1211int irdma_sc_ceq_init(struct irdma_sc_ceq *ceq,
1212		      struct irdma_ceq_init_info *info);
1213void irdma_sc_cleanup_ceqes(struct irdma_sc_cq *cq, struct irdma_sc_ceq *ceq);
1214void *irdma_sc_process_ceq(struct irdma_sc_dev *dev, struct irdma_sc_ceq *ceq);
1215
1216int irdma_sc_aeq_init(struct irdma_sc_aeq *aeq,
1217		      struct irdma_aeq_init_info *info);
1218int irdma_sc_get_next_aeqe(struct irdma_sc_aeq *aeq,
1219			   struct irdma_aeqe_info *info);
1220void irdma_sc_repost_aeq_entries(struct irdma_sc_dev *dev, u32 count);
1221
1222void irdma_sc_pd_init(struct irdma_sc_dev *dev, struct irdma_sc_pd *pd, u32 pd_id,
1223		      int abi_ver);
1224void irdma_cfg_aeq(struct irdma_sc_dev *dev, u32 idx, bool enable);
1225void irdma_check_cqp_progress(struct irdma_cqp_timeout *cqp_timeout,
1226			      struct irdma_sc_dev *dev);
1227int irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err, u16 *min_err);
1228int irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp, bool free_hwcqp);
1229int irdma_sc_cqp_init(struct irdma_sc_cqp *cqp,
1230		      struct irdma_cqp_init_info *info);
1231void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp);
1232int irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 opcode,
1233				  struct irdma_ccq_cqe_info *cmpl_info);
1234int irdma_sc_qp_create(struct irdma_sc_qp *qp,
1235		       struct irdma_create_qp_info *info, u64 scratch,
1236		       bool post_sq);
1237int irdma_sc_qp_destroy(struct irdma_sc_qp *qp, u64 scratch,
1238			bool remove_hash_idx, bool ignore_mw_bnd, bool post_sq);
1239int irdma_sc_qp_flush_wqes(struct irdma_sc_qp *qp,
1240			   struct irdma_qp_flush_info *info, u64 scratch,
1241			   bool post_sq);
1242int irdma_sc_qp_init(struct irdma_sc_qp *qp, struct irdma_qp_init_info *info);
1243int irdma_sc_qp_modify(struct irdma_sc_qp *qp,
1244		       struct irdma_modify_qp_info *info, u64 scratch,
1245		       bool post_sq);
1246void irdma_sc_send_lsmm(struct irdma_sc_qp *qp, void *lsmm_buf, u32 size,
1247			irdma_stag stag);
1248void irdma_sc_send_rtt(struct irdma_sc_qp *qp, bool read);
1249void irdma_sc_qp_setctx(struct irdma_sc_qp *qp, __le64 *qp_ctx,
1250			struct irdma_qp_host_ctx_info *info);
1251void irdma_sc_qp_setctx_roce(struct irdma_sc_qp *qp, __le64 *qp_ctx,
1252			     struct irdma_qp_host_ctx_info *info);
1253int irdma_sc_cq_destroy(struct irdma_sc_cq *cq, u64 scratch, bool post_sq);
1254int irdma_sc_cq_init(struct irdma_sc_cq *cq, struct irdma_cq_init_info *info);
1255void irdma_sc_cq_resize(struct irdma_sc_cq *cq, struct irdma_modify_cq_info *info);
1256int irdma_sc_aeq_destroy(struct irdma_sc_aeq *aeq, u64 scratch, bool post_sq);
1257int irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
1258					u16 hmc_fn_id, bool post_sq,
1259					bool poll_registers);
1260
1261void sc_vsi_update_stats(struct irdma_sc_vsi *vsi);
1262struct cqp_info {
1263	union {
1264		struct {
1265			struct irdma_sc_qp *qp;
1266			struct irdma_create_qp_info info;
1267			u64 scratch;
1268		} qp_create;
1269
1270		struct {
1271			struct irdma_sc_qp *qp;
1272			struct irdma_modify_qp_info info;
1273			u64 scratch;
1274		} qp_modify;
1275
1276		struct {
1277			struct irdma_sc_qp *qp;
1278			u64 scratch;
1279			bool remove_hash_idx;
1280			bool ignore_mw_bnd;
1281		} qp_destroy;
1282
1283		struct {
1284			struct irdma_sc_cq *cq;
1285			u64 scratch;
1286			bool check_overflow;
1287		} cq_create;
1288
1289		struct {
1290			struct irdma_sc_cq *cq;
1291			struct irdma_modify_cq_info info;
1292			u64 scratch;
1293		} cq_modify;
1294
1295		struct {
1296			struct irdma_sc_cq *cq;
1297			u64 scratch;
1298		} cq_destroy;
1299
1300		struct {
1301			struct irdma_sc_dev *dev;
1302			struct irdma_allocate_stag_info info;
1303			u64 scratch;
1304		} alloc_stag;
1305
1306		struct {
1307			struct irdma_sc_dev *dev;
1308			struct irdma_mw_alloc_info info;
1309			u64 scratch;
1310		} mw_alloc;
1311
1312		struct {
1313			struct irdma_sc_dev *dev;
1314			struct irdma_reg_ns_stag_info info;
1315			u64 scratch;
1316		} mr_reg_non_shared;
1317
1318		struct {
1319			struct irdma_sc_dev *dev;
1320			struct irdma_dealloc_stag_info info;
1321			u64 scratch;
1322		} dealloc_stag;
1323
1324		struct {
1325			struct irdma_sc_cqp *cqp;
1326			struct irdma_add_arp_cache_entry_info info;
1327			u64 scratch;
1328		} add_arp_cache_entry;
1329
1330		struct {
1331			struct irdma_sc_cqp *cqp;
1332			u64 scratch;
1333			u16 arp_index;
1334		} del_arp_cache_entry;
1335
1336		struct {
1337			struct irdma_sc_cqp *cqp;
1338			struct irdma_local_mac_entry_info info;
1339			u64 scratch;
1340		} add_local_mac_entry;
1341
1342		struct {
1343			struct irdma_sc_cqp *cqp;
1344			u64 scratch;
1345			u8 entry_idx;
1346			u8 ignore_ref_count;
1347		} del_local_mac_entry;
1348
1349		struct {
1350			struct irdma_sc_cqp *cqp;
1351			u64 scratch;
1352		} alloc_local_mac_entry;
1353
1354		struct {
1355			struct irdma_sc_cqp *cqp;
1356			struct irdma_cqp_manage_push_page_info info;
1357			u64 scratch;
1358		} manage_push_page;
1359
1360		struct {
1361			struct irdma_sc_dev *dev;
1362			struct irdma_upload_context_info info;
1363			u64 scratch;
1364		} qp_upload_context;
1365
1366		struct {
1367			struct irdma_sc_dev *dev;
1368			struct irdma_hmc_fcn_info info;
1369			u64 scratch;
1370		} manage_hmc_pm;
1371
1372		struct {
1373			struct irdma_sc_ceq *ceq;
1374			u64 scratch;
1375		} ceq_create;
1376
1377		struct {
1378			struct irdma_sc_ceq *ceq;
1379			u64 scratch;
1380		} ceq_destroy;
1381
1382		struct {
1383			struct irdma_sc_aeq *aeq;
1384			u64 scratch;
1385		} aeq_create;
1386
1387		struct {
1388			struct irdma_sc_aeq *aeq;
1389			u64 scratch;
1390		} aeq_destroy;
1391
1392		struct {
1393			struct irdma_sc_qp *qp;
1394			struct irdma_qp_flush_info info;
1395			u64 scratch;
1396		} qp_flush_wqes;
1397
1398		struct {
1399			struct irdma_sc_qp *qp;
1400			struct irdma_gen_ae_info info;
1401			u64 scratch;
1402		} gen_ae;
1403
1404		struct {
1405			struct irdma_sc_cqp *cqp;
1406			void *fpm_val_va;
1407			u64 fpm_val_pa;
1408			u16 hmc_fn_id;
1409			u64 scratch;
1410		} query_fpm_val;
1411
1412		struct {
1413			struct irdma_sc_cqp *cqp;
1414			void *fpm_val_va;
1415			u64 fpm_val_pa;
1416			u16 hmc_fn_id;
1417			u64 scratch;
1418		} commit_fpm_val;
1419
1420		struct {
1421			struct irdma_sc_cqp *cqp;
1422			struct irdma_apbvt_info info;
1423			u64 scratch;
1424		} manage_apbvt_entry;
1425
1426		struct {
1427			struct irdma_sc_cqp *cqp;
1428			struct irdma_qhash_table_info info;
1429			u64 scratch;
1430		} manage_qhash_table_entry;
1431
1432		struct {
1433			struct irdma_sc_dev *dev;
1434			struct irdma_update_sds_info info;
1435			u64 scratch;
1436		} update_pe_sds;
1437
1438		struct {
1439			struct irdma_sc_cqp *cqp;
1440			struct irdma_sc_qp *qp;
1441			u64 scratch;
1442		} suspend_resume;
1443
1444		struct {
1445			struct irdma_sc_cqp *cqp;
1446			struct irdma_ah_info info;
1447			u64 scratch;
1448		} ah_create;
1449
1450		struct {
1451			struct irdma_sc_cqp *cqp;
1452			struct irdma_ah_info info;
1453			u64 scratch;
1454		} ah_destroy;
1455
1456		struct {
1457			struct irdma_sc_cqp *cqp;
1458			struct irdma_mcast_grp_info info;
1459			u64 scratch;
1460		} mc_create;
1461
1462		struct {
1463			struct irdma_sc_cqp *cqp;
1464			struct irdma_mcast_grp_info info;
1465			u64 scratch;
1466		} mc_destroy;
1467
1468		struct {
1469			struct irdma_sc_cqp *cqp;
1470			struct irdma_mcast_grp_info info;
1471			u64 scratch;
1472		} mc_modify;
1473
1474		struct {
1475			struct irdma_sc_cqp *cqp;
1476			struct irdma_stats_inst_info info;
1477			u64 scratch;
1478		} stats_manage;
1479
1480		struct {
1481			struct irdma_sc_cqp *cqp;
1482			struct irdma_stats_gather_info info;
1483			u64 scratch;
1484		} stats_gather;
1485
1486		struct {
1487			struct irdma_sc_cqp *cqp;
1488			struct irdma_ws_node_info info;
1489			u64 scratch;
1490		} ws_node;
1491
1492		struct {
1493			struct irdma_sc_cqp *cqp;
1494			struct irdma_up_info info;
1495			u64 scratch;
1496		} up_map;
1497
1498		struct {
1499			struct irdma_sc_cqp *cqp;
1500			struct irdma_dma_mem query_buff_mem;
1501			u64 scratch;
1502		} query_rdma;
1503	} u;
1504};
1505
1506struct cqp_cmds_info {
1507	struct list_head cqp_cmd_entry;
1508	u8 cqp_cmd;
1509	u8 post_sq;
1510	struct cqp_info in;
1511};
1512
1513__le64 *irdma_sc_cqp_get_next_send_wqe_idx(struct irdma_sc_cqp *cqp, u64 scratch,
1514					   u32 *wqe_idx);
1515
1516/**
1517 * irdma_sc_cqp_get_next_send_wqe - get next wqe on cqp sq
1518 * @cqp: struct for cqp hw
1519 * @scratch: private data for CQP WQE
1520 */
1521static inline __le64 *irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch)
1522{
1523	u32 wqe_idx;
1524
1525	return irdma_sc_cqp_get_next_send_wqe_idx(cqp, scratch, &wqe_idx);
1526}
1527#endif /* IRDMA_TYPE_H */
1528