1/* 2 * linux/drivers/net/ehea/ehea_hw.h 3 * 4 * eHEA ethernet device driver for IBM eServer System p 5 * 6 * (C) Copyright IBM Corp. 2006 7 * 8 * Authors: 9 * Christoph Raisch <raisch@de.ibm.com> 10 * Jan-Bernd Themann <themann@de.ibm.com> 11 * Thomas Klein <tklein@de.ibm.com> 12 * 13 * 14 * This program is free software; you can redistribute it and/or modify 15 * it under the terms of the GNU General Public License as published by 16 * the Free Software Foundation; either version 2, or (at your option) 17 * any later version. 18 * 19 * This program is distributed in the hope that it will be useful, 20 * but WITHOUT ANY WARRANTY; without even the implied warranty of 21 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 22 * GNU General Public License for more details. 23 * 24 * You should have received a copy of the GNU General Public License 25 * along with this program; if not, write to the Free Software 26 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 27 */ 28 29#ifndef __EHEA_HW_H__ 30#define __EHEA_HW_H__ 31 32#define QPX_SQA_VALUE EHEA_BMASK_IBM(48, 63) 33#define QPX_RQ1A_VALUE EHEA_BMASK_IBM(48, 63) 34#define QPX_RQ2A_VALUE EHEA_BMASK_IBM(48, 63) 35#define QPX_RQ3A_VALUE EHEA_BMASK_IBM(48, 63) 36 37#define QPTEMM_OFFSET(x) offsetof(struct ehea_qptemm, x) 38 39struct ehea_qptemm { 40 u64 qpx_hcr; 41 u64 qpx_c; 42 u64 qpx_herr; 43 u64 qpx_aer; 44 u64 qpx_sqa; 45 u64 qpx_sqc; 46 u64 qpx_rq1a; 47 u64 qpx_rq1c; 48 u64 qpx_st; 49 u64 qpx_aerr; 50 u64 qpx_tenure; 51 u64 qpx_reserved1[(0x098 - 0x058) / 8]; 52 u64 qpx_portp; 53 u64 qpx_reserved2[(0x100 - 0x0A0) / 8]; 54 u64 qpx_t; 55 u64 qpx_sqhp; 56 u64 qpx_sqptp; 57 u64 qpx_reserved3[(0x140 - 0x118) / 8]; 58 u64 qpx_sqwsize; 59 u64 qpx_reserved4[(0x170 - 0x148) / 8]; 60 u64 qpx_sqsize; 61 u64 qpx_reserved5[(0x1B0 - 0x178) / 8]; 62 u64 qpx_sigt; 63 u64 qpx_wqecnt; 64 u64 qpx_rq1hp; 65 u64 qpx_rq1ptp; 66 u64 qpx_rq1size; 67 u64 qpx_reserved6[(0x220 - 0x1D8) / 8]; 68 u64 qpx_rq1wsize; 69 u64 qpx_reserved7[(0x240 - 0x228) / 8]; 70 u64 qpx_pd; 71 u64 qpx_scqn; 72 u64 qpx_rcqn; 73 u64 qpx_aeqn; 74 u64 reserved49; 75 u64 qpx_ram; 76 u64 qpx_reserved8[(0x300 - 0x270) / 8]; 77 u64 qpx_rq2a; 78 u64 qpx_rq2c; 79 u64 qpx_rq2hp; 80 u64 qpx_rq2ptp; 81 u64 qpx_rq2size; 82 u64 qpx_rq2wsize; 83 u64 qpx_rq2th; 84 u64 qpx_rq3a; 85 u64 qpx_rq3c; 86 u64 qpx_rq3hp; 87 u64 qpx_rq3ptp; 88 u64 qpx_rq3size; 89 u64 qpx_rq3wsize; 90 u64 qpx_rq3th; 91 u64 qpx_lpn; 92 u64 qpx_reserved9[(0x400 - 0x378) / 8]; 93 u64 reserved_ext[(0x500 - 0x400) / 8]; 94 u64 reserved2[(0x1000 - 0x500) / 8]; 95}; 96 97#define MRx_HCR_LPARID_VALID EHEA_BMASK_IBM(0, 0) 98 99#define MRMWMM_OFFSET(x) offsetof(struct ehea_mrmwmm, x) 100 101struct ehea_mrmwmm { 102 u64 mrx_hcr; 103 u64 mrx_c; 104 u64 mrx_herr; 105 u64 mrx_aer; 106 u64 mrx_pp; 107 u64 reserved1; 108 u64 reserved2; 109 u64 reserved3; 110 u64 reserved4[(0x200 - 0x40) / 8]; 111 u64 mrx_ctl[64]; 112}; 113 114#define QPEDMM_OFFSET(x) offsetof(struct ehea_qpedmm, x) 115 116struct ehea_qpedmm { 117 118 u64 reserved0[(0x400) / 8]; 119 u64 qpedx_phh; 120 u64 qpedx_ppsgp; 121 u64 qpedx_ppsgu; 122 u64 qpedx_ppdgp; 123 u64 qpedx_ppdgu; 124 u64 qpedx_aph; 125 u64 qpedx_apsgp; 126 u64 qpedx_apsgu; 127 u64 qpedx_apdgp; 128 u64 qpedx_apdgu; 129 u64 qpedx_apav; 130 u64 qpedx_apsav; 131 u64 qpedx_hcr; 132 u64 reserved1[4]; 133 u64 qpedx_rrl0; 134 u64 qpedx_rrrkey0; 135 u64 qpedx_rrva0; 136 u64 reserved2; 137 u64 qpedx_rrl1; 138 u64 qpedx_rrrkey1; 139 u64 qpedx_rrva1; 140 u64 reserved3; 141 u64 qpedx_rrl2; 142 u64 qpedx_rrrkey2; 143 u64 qpedx_rrva2; 144 u64 reserved4; 145 u64 qpedx_rrl3; 146 u64 qpedx_rrrkey3; 147 u64 qpedx_rrva3; 148}; 149 150#define CQX_FECADDER EHEA_BMASK_IBM(32, 63) 151#define CQX_FEC_CQE_CNT EHEA_BMASK_IBM(32, 63) 152#define CQX_N1_GENERATE_COMP_EVENT EHEA_BMASK_IBM(0, 0) 153#define CQX_EP_EVENT_PENDING EHEA_BMASK_IBM(0, 0) 154 155#define CQTEMM_OFFSET(x) offsetof(struct ehea_cqtemm, x) 156 157struct ehea_cqtemm { 158 u64 cqx_hcr; 159 u64 cqx_c; 160 u64 cqx_herr; 161 u64 cqx_aer; 162 u64 cqx_ptp; 163 u64 cqx_tp; 164 u64 cqx_fec; 165 u64 cqx_feca; 166 u64 cqx_ep; 167 u64 cqx_eq; 168 u64 reserved1; 169 u64 cqx_n0; 170 u64 cqx_n1; 171 u64 reserved2[(0x1000 - 0x60) / 8]; 172}; 173 174#define EQTEMM_OFFSET(x) offsetof(struct ehea_eqtemm, x) 175 176struct ehea_eqtemm { 177 u64 eqx_hcr; 178 u64 eqx_c; 179 u64 eqx_herr; 180 u64 eqx_aer; 181 u64 eqx_ptp; 182 u64 eqx_tp; 183 u64 eqx_ssba; 184 u64 eqx_psba; 185 u64 eqx_cec; 186 u64 eqx_meql; 187 u64 eqx_xisbi; 188 u64 eqx_xisc; 189 u64 eqx_it; 190}; 191 192/* 193 * These access functions will be changed when the dissuccsion about 194 * the new access methods for POWER has settled. 195 */ 196 197static inline u64 epa_load(struct h_epa epa, u32 offset) 198{ 199 return __raw_readq((void __iomem *)(epa.addr + offset)); 200} 201 202static inline void epa_store(struct h_epa epa, u32 offset, u64 value) 203{ 204 __raw_writeq(value, (void __iomem *)(epa.addr + offset)); 205 epa_load(epa, offset); /* synchronize explicitly to eHEA */ 206} 207 208static inline void epa_store_acc(struct h_epa epa, u32 offset, u64 value) 209{ 210 __raw_writeq(value, (void __iomem *)(epa.addr + offset)); 211} 212 213#define epa_store_eq(epa, offset, value)\ 214 epa_store(epa, EQTEMM_OFFSET(offset), value) 215#define epa_load_eq(epa, offset)\ 216 epa_load(epa, EQTEMM_OFFSET(offset)) 217 218#define epa_store_cq(epa, offset, value)\ 219 epa_store(epa, CQTEMM_OFFSET(offset), value) 220#define epa_load_cq(epa, offset)\ 221 epa_load(epa, CQTEMM_OFFSET(offset)) 222 223#define epa_store_qp(epa, offset, value)\ 224 epa_store(epa, QPTEMM_OFFSET(offset), value) 225#define epa_load_qp(epa, offset)\ 226 epa_load(epa, QPTEMM_OFFSET(offset)) 227 228#define epa_store_qped(epa, offset, value)\ 229 epa_store(epa, QPEDMM_OFFSET(offset), value) 230#define epa_load_qped(epa, offset)\ 231 epa_load(epa, QPEDMM_OFFSET(offset)) 232 233#define epa_store_mrmw(epa, offset, value)\ 234 epa_store(epa, MRMWMM_OFFSET(offset), value) 235#define epa_load_mrmw(epa, offset)\ 236 epa_load(epa, MRMWMM_OFFSET(offset)) 237 238#define epa_store_base(epa, offset, value)\ 239 epa_store(epa, HCAGR_OFFSET(offset), value) 240#define epa_load_base(epa, offset)\ 241 epa_load(epa, HCAGR_OFFSET(offset)) 242 243static inline void ehea_update_sqa(struct ehea_qp *qp, u16 nr_wqes) 244{ 245 struct h_epa epa = qp->epas.kernel; 246 epa_store_acc(epa, QPTEMM_OFFSET(qpx_sqa), 247 EHEA_BMASK_SET(QPX_SQA_VALUE, nr_wqes)); 248} 249 250static inline void ehea_update_rq3a(struct ehea_qp *qp, u16 nr_wqes) 251{ 252 struct h_epa epa = qp->epas.kernel; 253 epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq3a), 254 EHEA_BMASK_SET(QPX_RQ1A_VALUE, nr_wqes)); 255} 256 257static inline void ehea_update_rq2a(struct ehea_qp *qp, u16 nr_wqes) 258{ 259 struct h_epa epa = qp->epas.kernel; 260 epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq2a), 261 EHEA_BMASK_SET(QPX_RQ2A_VALUE, nr_wqes)); 262} 263 264static inline void ehea_update_rq1a(struct ehea_qp *qp, u16 nr_wqes) 265{ 266 struct h_epa epa = qp->epas.kernel; 267 epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq1a), 268 EHEA_BMASK_SET(QPX_RQ3A_VALUE, nr_wqes)); 269} 270 271static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes) 272{ 273 struct h_epa epa = cq->epas.kernel; 274 epa_store_acc(epa, CQTEMM_OFFSET(cqx_feca), 275 EHEA_BMASK_SET(CQX_FECADDER, nr_cqes)); 276} 277 278static inline void ehea_reset_cq_n1(struct ehea_cq *cq) 279{ 280 struct h_epa epa = cq->epas.kernel; 281 epa_store_cq(epa, cqx_n1, 282 EHEA_BMASK_SET(CQX_N1_GENERATE_COMP_EVENT, 1)); 283} 284 285static inline void ehea_reset_cq_ep(struct ehea_cq *my_cq) 286{ 287 struct h_epa epa = my_cq->epas.kernel; 288 epa_store_acc(epa, CQTEMM_OFFSET(cqx_ep), 289 EHEA_BMASK_SET(CQX_EP_EVENT_PENDING, 0)); 290} 291 292#endif /* __EHEA_HW_H__ */ 293