Lines Matching defs:ha

116 	struct qla_hw_data *ha;
124 ha = vha->hw;
138 if (!ha->max_qpairs) {
139 qpair = ha->base_qpair;
141 if (ha->queue_pair_map[qidx]) {
142 *handle = ha->queue_pair_map[qidx];
305 struct qla_hw_data *ha = fcport->vha->hw;
314 if (!ha->flags.fw_started || fcport->deleted == QLA_SESS_DELETED)
317 if (ha->flags.host_shutting_down) {
333 rval = ha->isp_ops->abort_command(sp);
369 struct qla_hw_data *ha = vha->hw;
379 if (!ha->flags.fw_started)
408 dma_sync_single_for_device(&ha->pdev->dev, nvme->u.nvme.cmd_dma,
441 qla_nvme_ls_reject_iocb(vha, ha->base_qpair, &a, true);
477 struct qla_hw_data *ha;
484 ha = vha->hw;
486 if (!ha->flags.fw_started)
512 dma_sync_single_for_device(&ha->pdev->dev, nvme->u.nvme.cmd_dma,
564 struct qla_hw_data *ha = vha->hw;
596 if (IS_SHADOW_REG_CAPABLE(ha)) {
620 atomic_inc(&ha->nvme_active_aen_cnt);
767 struct qla_hw_data *ha;
788 ha = vha->hw;
803 qpair = qla_mapq_nvme_select_qpair(ha, qpair);
937 struct qla_hw_data *ha;
944 ha = vha->hw;
952 } else if (ql2xnvme_queues > (ha->max_qpairs - 1)) {
955 ql2xnvme_queues, (ha->max_qpairs - 1),
956 (ha->max_qpairs - 1));
957 ql2xnvme_queues = ((ha->max_qpairs - 1));
962 (uint8_t)((ha->max_qpairs - 1) ? (ha->max_qpairs - 1) : 1));
973 mutex_lock(&ha->vport_lock);
985 get_device(&ha->pdev->dev),
987 mutex_unlock(&ha->vport_lock);
989 mutex_unlock(&ha->vport_lock);
1004 struct qla_hw_data *ha;
1009 ha = orig_sp->fcport->vha->hw;
1018 abt->drv.rsp_timeout = cpu_to_le16(2 * (ha->r_a_tov / 10));
1203 qla2xxx_get_vha_from_vp_idx(struct qla_hw_data *ha, uint16_t vp_index)
1208 base_vha = pci_get_drvdata(ha->pdev);
1210 if (!vp_index && !ha->num_vhosts)
1213 spin_lock_irqsave(&ha->vport_slock, flags);
1214 list_for_each_entry_safe(vha, tvp, &ha->vp_list, list) {
1216 spin_unlock_irqrestore(&ha->vport_slock, flags);
1220 spin_unlock_irqrestore(&ha->vport_slock, flags);
1233 struct qla_hw_data *ha;
1242 ha = rsp_q->hw;
1244 vha = qla2xxx_get_vha_from_vp_idx(ha, p->vp_index);