Lines Matching defs:devp

78 	struct c4iw_dev *devp;
337 qpd->devp = inode->i_private;
344 xa_for_each(&qpd->devp->qps, index, qp)
354 xa_lock_irq(&qpd->devp->qps);
355 xa_for_each(&qpd->devp->qps, index, qp)
357 xa_unlock_irq(&qpd->devp->qps);
383 ret = cxgb4_read_tpte(stagd->devp->rdev.lldi.ports[0], (u32)id<<8,
386 dev_err(&stagd->devp->rdev.lldi.pdev->dev,
432 stagd->devp = inode->i_private;
435 xa_for_each(&stagd->devp->mrs, index, p)
445 xa_lock_irq(&stagd->devp->mrs);
446 xa_for_each(&stagd->devp->mrs, index, p)
448 xa_unlock_irq(&stagd->devp->mrs);
683 epd->devp = inode->i_private;
686 xa_for_each(&epd->devp->hwtids, index, ep)
688 xa_for_each(&epd->devp->atids, index, ep)
690 xa_for_each(&epd->devp->stids, index, lep)
700 xa_lock_irq(&epd->devp->hwtids);
701 xa_for_each(&epd->devp->hwtids, index, ep)
703 xa_unlock_irq(&epd->devp->hwtids);
704 xa_lock_irq(&epd->devp->atids);
705 xa_for_each(&epd->devp->atids, index, ep)
707 xa_unlock_irq(&epd->devp->atids);
708 xa_lock_irq(&epd->devp->stids);
709 xa_for_each(&epd->devp->stids, index, lep)
711 xa_unlock_irq(&epd->devp->stids);
728 static void setup_debugfs(struct c4iw_dev *devp)
730 debugfs_create_file_size("qps", S_IWUSR, devp->debugfs_root,
731 (void *)devp, &qp_debugfs_fops, 4096);
733 debugfs_create_file_size("stags", S_IWUSR, devp->debugfs_root,
734 (void *)devp, &stag_debugfs_fops, 4096);
736 debugfs_create_file_size("stats", S_IWUSR, devp->debugfs_root,
737 (void *)devp, &stats_debugfs_fops, 4096);
739 debugfs_create_file_size("eps", S_IWUSR, devp->debugfs_root,
740 (void *)devp, &ep_debugfs_fops, 4096);
743 debugfs_create_file_size("wr_log", S_IWUSR, devp->debugfs_root,
744 (void *)devp, &wr_log_debugfs_fops, 4096);
970 struct c4iw_dev *devp;
982 devp = ib_alloc_device(c4iw_dev, ibdev);
983 if (!devp) {
987 devp->rdev.lldi = *infop;
991 devp->rdev.lldi.sge_ingpadboundary,
992 devp->rdev.lldi.sge_egrstatuspagesize);
994 devp->rdev.hw_queue.t4_eq_status_entries =
995 devp->rdev.lldi.sge_egrstatuspagesize / 64;
996 devp->rdev.hw_queue.t4_max_eq_size = 65520;
997 devp->rdev.hw_queue.t4_max_iq_size = 65520;
998 devp->rdev.hw_queue.t4_max_rq_size = 8192 -
999 devp->rdev.hw_queue.t4_eq_status_entries - 1;
1000 devp->rdev.hw_queue.t4_max_sq_size =
1001 devp->rdev.hw_queue.t4_max_eq_size -
1002 devp->rdev.hw_queue.t4_eq_status_entries - 1;
1003 devp->rdev.hw_queue.t4_max_qp_depth =
1004 devp->rdev.hw_queue.t4_max_rq_size;
1005 devp->rdev.hw_queue.t4_max_cq_depth =
1006 devp->rdev.hw_queue.t4_max_iq_size - 2;
1007 devp->rdev.hw_queue.t4_stat_len =
1008 devp->rdev.lldi.sge_egrstatuspagesize;
1015 devp->rdev.bar2_pa = pci_resource_start(devp->rdev.lldi.pdev, 2);
1016 if (!is_t4(devp->rdev.lldi.adapter_type)) {
1017 devp->rdev.bar2_kva = ioremap_wc(devp->rdev.bar2_pa,
1018 pci_resource_len(devp->rdev.lldi.pdev, 2));
1019 if (!devp->rdev.bar2_kva) {
1021 ib_dealloc_device(&devp->ibdev);
1025 devp->rdev.oc_mw_pa =
1026 pci_resource_start(devp->rdev.lldi.pdev, 2) +
1027 pci_resource_len(devp->rdev.lldi.pdev, 2) -
1028 roundup_pow_of_two(devp->rdev.lldi.vr->ocq.size);
1029 devp->rdev.oc_mw_kva = ioremap_wc(devp->rdev.oc_mw_pa,
1030 devp->rdev.lldi.vr->ocq.size);
1031 if (!devp->rdev.oc_mw_kva) {
1033 ib_dealloc_device(&devp->ibdev);
1039 devp->rdev.lldi.vr->ocq.start, devp->rdev.lldi.vr->ocq.size,
1040 devp->rdev.oc_mw_pa, devp->rdev.oc_mw_kva);
1042 ret = c4iw_rdev_open(&devp->rdev);
1045 ib_dealloc_device(&devp->ibdev);
1049 xa_init_flags(&devp->cqs, XA_FLAGS_LOCK_IRQ);
1050 xa_init_flags(&devp->qps, XA_FLAGS_LOCK_IRQ);
1051 xa_init_flags(&devp->mrs, XA_FLAGS_LOCK_IRQ);
1052 xa_init_flags(&devp->hwtids, XA_FLAGS_LOCK_IRQ);
1053 xa_init_flags(&devp->atids, XA_FLAGS_LOCK_IRQ);
1054 xa_init_flags(&devp->stids, XA_FLAGS_LOCK_IRQ);
1055 mutex_init(&devp->rdev.stats.lock);
1056 mutex_init(&devp->db_mutex);
1057 INIT_LIST_HEAD(&devp->db_fc_list);
1058 init_waitqueue_head(&devp->wait);
1059 devp->avail_ird = devp->rdev.lldi.max_ird_adapter;
1062 devp->debugfs_root = debugfs_create_dir(
1063 pci_name(devp->rdev.lldi.pdev),
1065 setup_debugfs(devp);
1069 return devp;