Lines Matching defs:ha

60 	struct qla_hw_data *ha = rsp->hw;
62 if (IS_P3P_TYPE(ha))
65 ha->isp_ops->intr_handler(0, rsp);
93 qla2x00_clean_dsd_pool(struct qla_hw_data *ha, struct crc_context *ctx)
99 dma_pool_free(ha->dl_dma_pool, dsd->dsd_addr,
256 qla2x00_gid_list_size(struct qla_hw_data *ha)
258 if (IS_QLAFX00(ha))
261 return sizeof(struct gid_list_info) * ha->max_fibre_devices;
265 qla2x00_handle_mbx_completion(struct qla_hw_data *ha, int status)
267 if (test_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags) &&
268 (status & MBX_INTERRUPT) && ha->flags.mbox_int) {
269 set_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags);
270 clear_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags);
271 complete(&ha->mbx_intr_comp);
348 for (i = 0; i < tgt->ha->max_qpairs + 1; i++) {
397 struct qla_hw_data *ha = qp->hw;
408 iocbs_used = ha->base_qpair->fwres.iocbs_used;
409 for (i = 0; i < ha->max_qpairs; i++) {
410 if (ha->queue_pair_map[i])
411 iocbs_used += ha->queue_pair_map[i]->fwres.iocbs_used;
421 exch_used = ha->base_qpair->fwres.exch_used;
422 for (i = 0; i < ha->max_qpairs; i++) {
423 if (ha->queue_pair_map[i])
424 exch_used += ha->queue_pair_map[i]->fwres.exch_used;
434 if ((iores->iocb_cnt + atomic_read(&ha->fwres.iocb_used)) >=
435 ha->fwres.iocb_limit) {
441 if ((iores->exch_cnt + atomic_read(&ha->fwres.exch_used)) >=
442 ha->fwres.exch_limit) {
453 atomic_add(iores->iocb_cnt, &ha->fwres.iocb_used);
454 atomic_add(iores->exch_cnt, &ha->fwres.exch_used);
485 struct qla_hw_data *ha = qp->hw;
489 qla_atomic_dtz(&ha->fwres.iocb_used, iores->iocb_cnt);
492 qla_atomic_dtz(&ha->fwres.exch_used, iores->exch_cnt);
523 * ha = pointer to host adapter structure.
532 uint32_t qla2x00_isp_reg_stat(struct qla_hw_data *ha)
534 struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;
535 struct device_reg_82xx __iomem *reg82 = &ha->iobase->isp82;
537 if (IS_P3P_TYPE(ha))
578 qla_mapq_nvme_select_qpair(struct qla_hw_data *ha, struct qla_qpair *qpair)
583 ha->qp_cpu_map[cpuid]) {
584 qpair = ha->qp_cpu_map[cpuid];
590 qla_mapq_init_qp_cpu_map(struct qla_hw_data *ha,
597 if (!ha->qp_cpu_map)
599 mask = pci_irq_get_affinity(ha->pdev, msix->vector_base0);
604 ha->qp_cpu_map[cpu] = qpair;
611 qla_mapq_free_qp_cpu_map(struct qla_hw_data *ha)
613 if (ha->qp_cpu_map) {
614 kfree(ha->qp_cpu_map);
615 ha->qp_cpu_map = NULL;
619 static inline int qla_mapq_alloc_qp_cpu_map(struct qla_hw_data *ha)
621 scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
623 if (!ha->qp_cpu_map) {
624 ha->qp_cpu_map = kcalloc(NR_CPUS, sizeof(struct qla_qpair *),
626 if (!ha->qp_cpu_map) {