1/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
2/*
3 * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved.
4 * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved.
5 */
6
7#ifndef RXE_H
8#define RXE_H
9
10#ifdef pr_fmt
11#undef pr_fmt
12#endif
13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15#include <linux/skbuff.h>
16
17#include <rdma/ib_verbs.h>
18#include <rdma/ib_user_verbs.h>
19#include <rdma/ib_pack.h>
20#include <rdma/ib_smi.h>
21#include <rdma/ib_umem.h>
22#include <rdma/ib_cache.h>
23#include <rdma/ib_addr.h>
24#include <crypto/hash.h>
25
26#include "rxe_net.h"
27#include "rxe_opcode.h"
28#include "rxe_hdr.h"
29#include "rxe_param.h"
30#include "rxe_verbs.h"
31#include "rxe_loc.h"
32
33/*
34 * Version 1 and Version 2 are identical on 64 bit machines, but on 32 bit
35 * machines Version 2 has a different struct layout.
36 */
37#define RXE_UVERBS_ABI_VERSION		2
38
39#define RXE_ROCE_V2_SPORT		(0xc000)
40
41#define rxe_dbg(fmt, ...) pr_debug("%s: " fmt, __func__, ##__VA_ARGS__)
42#define rxe_dbg_dev(rxe, fmt, ...) ibdev_dbg(&(rxe)->ib_dev,		\
43		"%s: " fmt, __func__, ##__VA_ARGS__)
44#define rxe_dbg_uc(uc, fmt, ...) ibdev_dbg((uc)->ibuc.device,		\
45		"uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__)
46#define rxe_dbg_pd(pd, fmt, ...) ibdev_dbg((pd)->ibpd.device,		\
47		"pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__)
48#define rxe_dbg_ah(ah, fmt, ...) ibdev_dbg((ah)->ibah.device,		\
49		"ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__)
50#define rxe_dbg_srq(srq, fmt, ...) ibdev_dbg((srq)->ibsrq.device,	\
51		"srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__)
52#define rxe_dbg_qp(qp, fmt, ...) ibdev_dbg((qp)->ibqp.device,		\
53		"qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
54#define rxe_dbg_cq(cq, fmt, ...) ibdev_dbg((cq)->ibcq.device,		\
55		"cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__)
56#define rxe_dbg_mr(mr, fmt, ...) ibdev_dbg((mr)->ibmr.device,		\
57		"mr#%d %s:  " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__)
58#define rxe_dbg_mw(mw, fmt, ...) ibdev_dbg((mw)->ibmw.device,		\
59		"mw#%d %s:  " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
60
61#define rxe_err(fmt, ...) pr_err_ratelimited("%s: " fmt, __func__, \
62					##__VA_ARGS__)
63#define rxe_err_dev(rxe, fmt, ...) ibdev_err_ratelimited(&(rxe)->ib_dev, \
64		"%s: " fmt, __func__, ##__VA_ARGS__)
65#define rxe_err_uc(uc, fmt, ...) ibdev_err_ratelimited((uc)->ibuc.device, \
66		"uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__)
67#define rxe_err_pd(pd, fmt, ...) ibdev_err_ratelimited((pd)->ibpd.device, \
68		"pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__)
69#define rxe_err_ah(ah, fmt, ...) ibdev_err_ratelimited((ah)->ibah.device, \
70		"ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__)
71#define rxe_err_srq(srq, fmt, ...) ibdev_err_ratelimited((srq)->ibsrq.device, \
72		"srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__)
73#define rxe_err_qp(qp, fmt, ...) ibdev_err_ratelimited((qp)->ibqp.device, \
74		"qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
75#define rxe_err_cq(cq, fmt, ...) ibdev_err_ratelimited((cq)->ibcq.device, \
76		"cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__)
77#define rxe_err_mr(mr, fmt, ...) ibdev_err_ratelimited((mr)->ibmr.device, \
78		"mr#%d %s:  " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__)
79#define rxe_err_mw(mw, fmt, ...) ibdev_err_ratelimited((mw)->ibmw.device, \
80		"mw#%d %s:  " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
81
82#define rxe_info(fmt, ...) pr_info_ratelimited("%s: " fmt, __func__, \
83					##__VA_ARGS__)
84#define rxe_info_dev(rxe, fmt, ...) ibdev_info_ratelimited(&(rxe)->ib_dev, \
85		"%s: " fmt, __func__, ##__VA_ARGS__)
86#define rxe_info_uc(uc, fmt, ...) ibdev_info_ratelimited((uc)->ibuc.device, \
87		"uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__)
88#define rxe_info_pd(pd, fmt, ...) ibdev_info_ratelimited((pd)->ibpd.device, \
89		"pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__)
90#define rxe_info_ah(ah, fmt, ...) ibdev_info_ratelimited((ah)->ibah.device, \
91		"ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__)
92#define rxe_info_srq(srq, fmt, ...) ibdev_info_ratelimited((srq)->ibsrq.device, \
93		"srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__)
94#define rxe_info_qp(qp, fmt, ...) ibdev_info_ratelimited((qp)->ibqp.device, \
95		"qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
96#define rxe_info_cq(cq, fmt, ...) ibdev_info_ratelimited((cq)->ibcq.device, \
97		"cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__)
98#define rxe_info_mr(mr, fmt, ...) ibdev_info_ratelimited((mr)->ibmr.device, \
99		"mr#%d %s:  " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__)
100#define rxe_info_mw(mw, fmt, ...) ibdev_info_ratelimited((mw)->ibmw.device, \
101		"mw#%d %s:  " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
102
103/* responder states */
104enum resp_states {
105	RESPST_NONE,
106	RESPST_GET_REQ,
107	RESPST_CHK_PSN,
108	RESPST_CHK_OP_SEQ,
109	RESPST_CHK_OP_VALID,
110	RESPST_CHK_RESOURCE,
111	RESPST_CHK_LENGTH,
112	RESPST_CHK_RKEY,
113	RESPST_EXECUTE,
114	RESPST_READ_REPLY,
115	RESPST_ATOMIC_REPLY,
116	RESPST_ATOMIC_WRITE_REPLY,
117	RESPST_PROCESS_FLUSH,
118	RESPST_COMPLETE,
119	RESPST_ACKNOWLEDGE,
120	RESPST_CLEANUP,
121	RESPST_DUPLICATE_REQUEST,
122	RESPST_ERR_MALFORMED_WQE,
123	RESPST_ERR_UNSUPPORTED_OPCODE,
124	RESPST_ERR_MISALIGNED_ATOMIC,
125	RESPST_ERR_PSN_OUT_OF_SEQ,
126	RESPST_ERR_MISSING_OPCODE_FIRST,
127	RESPST_ERR_MISSING_OPCODE_LAST_C,
128	RESPST_ERR_MISSING_OPCODE_LAST_D1E,
129	RESPST_ERR_TOO_MANY_RDMA_ATM_REQ,
130	RESPST_ERR_RNR,
131	RESPST_ERR_RKEY_VIOLATION,
132	RESPST_ERR_INVALIDATE_RKEY,
133	RESPST_ERR_LENGTH,
134	RESPST_ERR_CQ_OVERFLOW,
135	RESPST_ERROR,
136	RESPST_DONE,
137	RESPST_EXIT,
138};
139
140void rxe_set_mtu(struct rxe_dev *rxe, unsigned int dev_mtu);
141
142int rxe_add(struct rxe_dev *rxe, unsigned int mtu, const char *ibdev_name);
143
144void rxe_rcv(struct sk_buff *skb);
145
146/* The caller must do a matching ib_device_put(&dev->ib_dev) */
147static inline struct rxe_dev *rxe_get_dev_from_net(struct net_device *ndev)
148{
149	struct ib_device *ibdev =
150		ib_device_get_by_netdev(ndev, RDMA_DRIVER_RXE);
151
152	if (!ibdev)
153		return NULL;
154	return container_of(ibdev, struct rxe_dev, ib_dev);
155}
156
157void rxe_port_up(struct rxe_dev *rxe);
158void rxe_port_down(struct rxe_dev *rxe);
159void rxe_set_port_state(struct rxe_dev *rxe);
160
161#endif /* RXE_H */
162