1/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */ 2/* 3 * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved. 4 * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved. 5 */ 6 7#ifndef RXE_H 8#define RXE_H 9 10#ifdef pr_fmt 11#undef pr_fmt 12#endif 13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 14 15#include <linux/skbuff.h> 16 17#include <rdma/ib_verbs.h> 18#include <rdma/ib_user_verbs.h> 19#include <rdma/ib_pack.h> 20#include <rdma/ib_smi.h> 21#include <rdma/ib_umem.h> 22#include <rdma/ib_cache.h> 23#include <rdma/ib_addr.h> 24#include <crypto/hash.h> 25 26#include "rxe_net.h" 27#include "rxe_opcode.h" 28#include "rxe_hdr.h" 29#include "rxe_param.h" 30#include "rxe_verbs.h" 31#include "rxe_loc.h" 32 33/* 34 * Version 1 and Version 2 are identical on 64 bit machines, but on 32 bit 35 * machines Version 2 has a different struct layout. 36 */ 37#define RXE_UVERBS_ABI_VERSION 2 38 39#define RXE_ROCE_V2_SPORT (0xc000) 40 41#define rxe_dbg(fmt, ...) pr_debug("%s: " fmt, __func__, ##__VA_ARGS__) 42#define rxe_dbg_dev(rxe, fmt, ...) ibdev_dbg(&(rxe)->ib_dev, \ 43 "%s: " fmt, __func__, ##__VA_ARGS__) 44#define rxe_dbg_uc(uc, fmt, ...) ibdev_dbg((uc)->ibuc.device, \ 45 "uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__) 46#define rxe_dbg_pd(pd, fmt, ...) ibdev_dbg((pd)->ibpd.device, \ 47 "pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__) 48#define rxe_dbg_ah(ah, fmt, ...) ibdev_dbg((ah)->ibah.device, \ 49 "ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__) 50#define rxe_dbg_srq(srq, fmt, ...) ibdev_dbg((srq)->ibsrq.device, \ 51 "srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__) 52#define rxe_dbg_qp(qp, fmt, ...) ibdev_dbg((qp)->ibqp.device, \ 53 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) 54#define rxe_dbg_cq(cq, fmt, ...) ibdev_dbg((cq)->ibcq.device, \ 55 "cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__) 56#define rxe_dbg_mr(mr, fmt, ...) ibdev_dbg((mr)->ibmr.device, \ 57 "mr#%d %s: " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__) 58#define rxe_dbg_mw(mw, fmt, ...) ibdev_dbg((mw)->ibmw.device, \ 59 "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__) 60 61#define rxe_err(fmt, ...) pr_err_ratelimited("%s: " fmt, __func__, \ 62 ##__VA_ARGS__) 63#define rxe_err_dev(rxe, fmt, ...) ibdev_err_ratelimited(&(rxe)->ib_dev, \ 64 "%s: " fmt, __func__, ##__VA_ARGS__) 65#define rxe_err_uc(uc, fmt, ...) ibdev_err_ratelimited((uc)->ibuc.device, \ 66 "uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__) 67#define rxe_err_pd(pd, fmt, ...) ibdev_err_ratelimited((pd)->ibpd.device, \ 68 "pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__) 69#define rxe_err_ah(ah, fmt, ...) ibdev_err_ratelimited((ah)->ibah.device, \ 70 "ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__) 71#define rxe_err_srq(srq, fmt, ...) ibdev_err_ratelimited((srq)->ibsrq.device, \ 72 "srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__) 73#define rxe_err_qp(qp, fmt, ...) ibdev_err_ratelimited((qp)->ibqp.device, \ 74 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) 75#define rxe_err_cq(cq, fmt, ...) ibdev_err_ratelimited((cq)->ibcq.device, \ 76 "cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__) 77#define rxe_err_mr(mr, fmt, ...) ibdev_err_ratelimited((mr)->ibmr.device, \ 78 "mr#%d %s: " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__) 79#define rxe_err_mw(mw, fmt, ...) ibdev_err_ratelimited((mw)->ibmw.device, \ 80 "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__) 81 82#define rxe_info(fmt, ...) pr_info_ratelimited("%s: " fmt, __func__, \ 83 ##__VA_ARGS__) 84#define rxe_info_dev(rxe, fmt, ...) ibdev_info_ratelimited(&(rxe)->ib_dev, \ 85 "%s: " fmt, __func__, ##__VA_ARGS__) 86#define rxe_info_uc(uc, fmt, ...) ibdev_info_ratelimited((uc)->ibuc.device, \ 87 "uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__) 88#define rxe_info_pd(pd, fmt, ...) ibdev_info_ratelimited((pd)->ibpd.device, \ 89 "pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__) 90#define rxe_info_ah(ah, fmt, ...) ibdev_info_ratelimited((ah)->ibah.device, \ 91 "ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__) 92#define rxe_info_srq(srq, fmt, ...) ibdev_info_ratelimited((srq)->ibsrq.device, \ 93 "srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__) 94#define rxe_info_qp(qp, fmt, ...) ibdev_info_ratelimited((qp)->ibqp.device, \ 95 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) 96#define rxe_info_cq(cq, fmt, ...) ibdev_info_ratelimited((cq)->ibcq.device, \ 97 "cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__) 98#define rxe_info_mr(mr, fmt, ...) ibdev_info_ratelimited((mr)->ibmr.device, \ 99 "mr#%d %s: " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__) 100#define rxe_info_mw(mw, fmt, ...) ibdev_info_ratelimited((mw)->ibmw.device, \ 101 "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__) 102 103/* responder states */ 104enum resp_states { 105 RESPST_NONE, 106 RESPST_GET_REQ, 107 RESPST_CHK_PSN, 108 RESPST_CHK_OP_SEQ, 109 RESPST_CHK_OP_VALID, 110 RESPST_CHK_RESOURCE, 111 RESPST_CHK_LENGTH, 112 RESPST_CHK_RKEY, 113 RESPST_EXECUTE, 114 RESPST_READ_REPLY, 115 RESPST_ATOMIC_REPLY, 116 RESPST_ATOMIC_WRITE_REPLY, 117 RESPST_PROCESS_FLUSH, 118 RESPST_COMPLETE, 119 RESPST_ACKNOWLEDGE, 120 RESPST_CLEANUP, 121 RESPST_DUPLICATE_REQUEST, 122 RESPST_ERR_MALFORMED_WQE, 123 RESPST_ERR_UNSUPPORTED_OPCODE, 124 RESPST_ERR_MISALIGNED_ATOMIC, 125 RESPST_ERR_PSN_OUT_OF_SEQ, 126 RESPST_ERR_MISSING_OPCODE_FIRST, 127 RESPST_ERR_MISSING_OPCODE_LAST_C, 128 RESPST_ERR_MISSING_OPCODE_LAST_D1E, 129 RESPST_ERR_TOO_MANY_RDMA_ATM_REQ, 130 RESPST_ERR_RNR, 131 RESPST_ERR_RKEY_VIOLATION, 132 RESPST_ERR_INVALIDATE_RKEY, 133 RESPST_ERR_LENGTH, 134 RESPST_ERR_CQ_OVERFLOW, 135 RESPST_ERROR, 136 RESPST_DONE, 137 RESPST_EXIT, 138}; 139 140void rxe_set_mtu(struct rxe_dev *rxe, unsigned int dev_mtu); 141 142int rxe_add(struct rxe_dev *rxe, unsigned int mtu, const char *ibdev_name); 143 144void rxe_rcv(struct sk_buff *skb); 145 146/* The caller must do a matching ib_device_put(&dev->ib_dev) */ 147static inline struct rxe_dev *rxe_get_dev_from_net(struct net_device *ndev) 148{ 149 struct ib_device *ibdev = 150 ib_device_get_by_netdev(ndev, RDMA_DRIVER_RXE); 151 152 if (!ibdev) 153 return NULL; 154 return container_of(ibdev, struct rxe_dev, ib_dev); 155} 156 157void rxe_port_up(struct rxe_dev *rxe); 158void rxe_port_down(struct rxe_dev *rxe); 159void rxe_set_port_state(struct rxe_dev *rxe); 160 161#endif /* RXE_H */ 162