/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | mad.c | 63 struct rvt_ibport *rvp; local 68 rvp = rdi->ports[p]; 78 rvp->send_agent = agent; 88 rvp = rdi->ports[p]; 89 if (rvp->send_agent) { 90 agent = rvp->send_agent; 91 rvp->send_agent = NULL; 110 struct rvt_ibport *rvp; local 114 rvp = rdi->ports[p]; 115 if (rvp [all...] |
H A D | vt.c | 123 struct rvt_ibport *rvp; local 126 rvp = rdi->ports[port_index]; 128 props->sm_lid = rvp->sm_lid; 129 props->sm_sl = rvp->sm_sl; 130 props->port_cap_flags = rvp->port_cap_flags; 133 props->bad_pkey_cntr = rvp->pkey_violations; 134 props->qkey_viol_cntr = rvp->qkey_violations; 135 props->subnet_timeout = rvp->subnet_timeout; 155 struct rvt_ibport *rvp; local 159 rvp 221 struct rvt_ibport *rvp; local [all...] |
H A D | qp.c | 731 struct rvt_ibport *rvp = rdi->ports[qp->port_num - 1]; local 738 if (rcu_dereference_protected(rvp->qp[0], 740 RCU_INIT_POINTER(rvp->qp[0], NULL); 741 } else if (rcu_dereference_protected(rvp->qp[1], 743 RCU_INIT_POINTER(rvp->qp[1], NULL); 1374 struct rvt_ibport *rvp = rdi->ports[qp->port_num - 1]; local 1381 rcu_assign_pointer(rvp->qp[qp->ibqp.qp_num], qp); 2596 struct rvt_ibport *rvp = rdi->ports[qp->port_num - 1]; local 2599 rvp->n_rc_timeouts++; 2709 struct rvt_ibport *rvp; local 2880 loopback_qp_drop(struct rvt_ibport *rvp, struct rvt_qp *sqp) argument 2904 struct rvt_ibport *rvp = NULL; local [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_mad.c | 73 agent = ibp->rvp.send_agent; 82 if (ibp->rvp.trap_timeout && 83 time_before(jiffies, ibp->rvp.trap_timeout)) 97 ibp->rvp.tid++; 98 smp->tid = cpu_to_be64(ibp->rvp.tid); 103 spin_lock_irqsave(&ibp->rvp.lock, flags); 104 if (!ibp->rvp.sm_ah) { 105 if (ibp->rvp.sm_lid != be16_to_cpu(IB_LID_PERMISSIVE)) { 108 ah = qib_create_qp0_ah(ibp, (u16)ibp->rvp.sm_lid); 113 ibp->rvp [all...] |
H A D | qib_sysfs.c | 406 static_assert(__same_type(((struct qib_ibport *)0)->rvp.n_##N, u64)); \ 410 offsetof(struct qib_ibport, rvp.n_##N) / sizeof(u64) \ 461 get_all_cpu_total(qibp->rvp.rc_acks) - 462 qibp->rvp.z_rc_acks); 472 return qib_store_per_cpu(dd, buf, count, &qibp->rvp.z_rc_acks, 473 get_all_cpu_total(qibp->rvp.rc_acks)); 484 get_all_cpu_total(qibp->rvp.rc_qacks) - 485 qibp->rvp.z_rc_qacks); 495 return qib_store_per_cpu(dd, buf, count, &qibp->rvp.z_rc_qacks, 496 get_all_cpu_total(qibp->rvp [all...] |
H A D | qib_ud.c | 67 qp = rvt_lookup_qpn(rdi, &ibp->rvp, rvt_get_swqe_remote_qpn(swqe)); 69 ibp->rvp.n_pkt_drops++; 80 ibp->rvp.n_pkt_drops++; 150 ibp->rvp.n_vl15_dropped++; 157 ibp->rvp.n_pkt_drops++; 215 ibp->rvp.n_loop_pkts++; 540 ibp->rvp.n_vl15_dropped++; 582 ibp->rvp.n_pkt_drops++;
|
H A D | qib_verbs.c | 231 ibp->rvp.n_pkt_drops++; 327 mcast = rvt_mcast_find(&ibp->rvp, &hdr->u.l.grh.dgid, lid); 343 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 355 ibp->rvp.n_pkt_drops++; 862 ibp->rvp.n_unaligned++; 1287 static int qib_get_guid_be(struct rvt_dev_info *rdi, struct rvt_ibport *rvp, argument 1290 struct qib_ibport *ibp = container_of(rvp, struct qib_ibport, rvp); 1352 qp0 = rcu_dereference(ibp->rvp.qp[0]); 1393 spin_lock_init(&ibp->rvp [all...] |
H A D | qib_ruc.c | 102 ibp->rvp.gid_prefix, guid)) 141 ibp->rvp.gid_prefix, guid)) 195 hdr->sgid.global.subnet_prefix = ibp->rvp.gid_prefix;
|
H A D | qib_init.c | 241 ppd->ibport_data.rvp.rc_acks = alloc_percpu(u64); 242 ppd->ibport_data.rvp.rc_qacks = alloc_percpu(u64); 243 ppd->ibport_data.rvp.rc_delayed_comp = alloc_percpu(u64); 244 if (!(ppd->ibport_data.rvp.rc_acks) || 245 !(ppd->ibport_data.rvp.rc_qacks) || 246 !(ppd->ibport_data.rvp.rc_delayed_comp)) 611 free_percpu(ppd->ibport_data.rvp.rc_acks); 612 free_percpu(ppd->ibport_data.rvp.rc_qacks); 613 free_percpu(ppd->ibport_data.rvp.rc_delayed_comp);
|
H A D | qib_rc.c | 713 this_cpu_inc(*ibp->rvp.rc_qacks); 841 ibp->rvp.n_rc_resends++; 843 ibp->rvp.n_rc_resends += (qp->s_psn - psn) & QIB_PSN_MASK; 967 this_cpu_inc(*ibp->rvp.rc_delayed_comp); 1106 this_cpu_inc(*ibp->rvp.rc_acks); 1138 ibp->rvp.n_rnr_naks++; 1153 ibp->rvp.n_rc_resends += (qp->s_psn - psn) & QIB_PSN_MASK; 1170 ibp->rvp.n_seq_naks++; 1183 ibp->rvp.n_other_naks++; 1188 ibp->rvp [all...] |
H A D | qib_driver.c | 352 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 367 ibp->rvp.n_pkt_drops++; 387 ibp->rvp.n_rc_seqnak++;
|
H A D | qib_qp.c | 224 if (rcu_dereference(ibp->rvp.qp[0])) 226 if (rcu_dereference(ibp->rvp.qp[1]))
|
H A D | qib_verbs.h | 145 struct rvt_ibport rvp; member in struct:qib_ibport
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | mad.c | 91 spin_lock_irqsave(&ibp->rvp.lock, flags); 92 list_replace_init(&ibp->rvp.trap_lists[i].list, &trap_list); 93 ibp->rvp.trap_lists[i].list_len = 0; 94 spin_unlock_irqrestore(&ibp->rvp.lock, flags); 138 spin_lock_irqsave(&ibp->rvp.lock, flags); 139 trap_list = &ibp->rvp.trap_lists[queue_id]; 166 if (!timer_pending(&ibp->rvp.trap_timer)) { 174 (1UL << ibp->rvp.subnet_timeout)) / 1000; 175 mod_timer(&ibp->rvp.trap_timer, 181 spin_unlock_irqrestore(&ibp->rvp [all...] |
H A D | ipoib_main.c | 61 qp = rvt_lookup_qpn(ib_to_rvt(priv->device), &ibp->rvp, qpn); 116 qp = rvt_lookup_qpn(ib_to_rvt(priv->device), &ibp->rvp, qpn); 146 qp = rvt_lookup_qpn(ib_to_rvt(priv->device), &ibp->rvp, qpn);
|
H A D | verbs.c | 360 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 378 ibp->rvp.n_pkt_drops++; 414 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 432 ibp->rvp.n_pkt_drops++; 472 mcast = rvt_mcast_find(&ibp->rvp, 487 ibp->rvp.n_pkt_drops++; 506 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 518 ibp->rvp.n_pkt_drops++; 526 ibp->rvp.n_pkt_drops++; 1456 static int hfi1_get_guid_be(struct rvt_dev_info *rdi, struct rvt_ibport *rvp, argument [all...] |
H A D | ud.c | 47 qp = rvt_lookup_qpn(ib_to_rvt(sqp->ibqp.device), &ibp->rvp, 50 ibp->rvp.n_pkt_drops++; 62 ibp->rvp.n_pkt_drops++; 132 ibp->rvp.n_vl15_dropped++; 139 ibp->rvp.n_pkt_drops++; 217 ibp->rvp.n_loop_pkts++; 786 if (ibp->rvp.port_cap_flags & IB_PORT_SM) 952 ibp->rvp.n_vl15_dropped++; 1022 ibp->rvp.n_pkt_drops++;
|
H A D | ruc.c | 53 if (!gid_ok(&packet->grh->dgid, ibp->rvp.gid_prefix, 90 if (!gid_ok(&packet->grh->dgid, ibp->rvp.gid_prefix, 138 hdr->sgid.global.subnet_prefix = ibp->rvp.gid_prefix;
|
H A D | hfi.h | 2372 ppd->ibport_data.rvp.z_rc_acks = 2373 get_all_cpu_total(ppd->ibport_data.rvp.rc_acks); 2374 ppd->ibport_data.rvp.z_rc_qacks = 2375 get_all_cpu_total(ppd->ibport_data.rvp.rc_qacks); 2451 rdma_ah_set_subnet_prefix(attr, ibp->rvp.gid_prefix); 2582 grh->sgid.global.subnet_prefix = ibp->rvp.gid_prefix; 2596 grh->dgid.global.subnet_prefix = ibp->rvp.gid_prefix;
|
H A D | driver.c | 270 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 285 ibp->rvp.n_pkt_drops++; 331 qp = rvt_lookup_qpn(rdi, &ibp->rvp, lqpn); 617 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qpn); 1518 ibp->rvp.n_pkt_drops++; 1601 ibp->rvp.n_pkt_drops++; 1703 ++ibp->rvp.n_pkt_drops;
|
H A D | rc.c | 1206 this_cpu_inc(*ibp->rvp.rc_qacks); 1618 ibp->rvp.n_rc_resends++; 1620 ibp->rvp.n_rc_resends += delta_psn(qp->s_psn, psn); 1848 this_cpu_inc(*ibp->rvp.rc_delayed_comp); 2073 this_cpu_inc(*ibp->rvp.rc_acks); 2163 ibp->rvp.n_rnr_naks++; 2192 ibp->rvp.n_rc_resends += delta_psn(qp->s_psn, psn); 2206 ibp->rvp.n_seq_naks++; 2219 ibp->rvp.n_other_naks++; 2224 ibp->rvp [all...] |
H A D | qp.c | 468 ibp->rvp.n_dmawait++; 722 if (rcu_dereference(ibp->rvp.qp[0])) 724 if (rcu_dereference(ibp->rvp.qp[1]))
|
/linux-master/include/linux/sunrpc/ |
H A D | cache.h | 307 static inline int get_expiry(char **bpp, time64_t *rvp) argument 312 error = get_time(bpp, rvp); 317 (*rvp) -= boot.tv_sec;
|
/linux-master/include/rdma/ |
H A D | rdmavt_qp.h | 700 struct rvt_ibport *rvp, 706 qp = rcu_dereference(rvp->qp[qpn]); 987 struct rvt_ibport *rvp = rvt_to_iport(qp); local 990 rvp->n_rc_crwaits++;
|
H A D | rdma_vt.h | 306 int (*get_guid_be)(struct rvt_dev_info *rdi, struct rvt_ibport *rvp,
|