Lines Matching defs:hdl

118  * irdma_find_handler - obtain hdl object to identify pf
124 struct irdma_handler *hdl;
128 list_for_each_entry(hdl, &irdma_handlers, list) {
129 if (!hdl->iwdev->rf->peer_info)
131 if (hdl->iwdev->rf->peer_info->dev == p_dev->dev) {
133 return hdl;
148 struct irdma_handler *hdl;
150 hdl = irdma_find_handler(peer);
151 if (!hdl) {
156 return hdl->iwdev;
562 struct irdma_handler *hdl;
569 hdl = irdma_find_handler(peer);
570 if (hdl)
573 hdl = kzalloc(sizeof(*hdl), GFP_KERNEL);
574 if (!hdl)
579 kfree(hdl);
586 kfree(hdl);
589 hdl->iwdev = iwdev;
590 iwdev->hdl = hdl;
599 irdma_add_handler(hdl);
610 TASK_INIT(&hdl->deferred_task, 0, irdma_finalize_task, &rf->dev_ctx.task_arg);
611 hdl->deferred_tq = taskqueue_create_fast("irdma_defer",
613 &hdl->deferred_tq);
614 taskqueue_start_threads(&hdl->deferred_tq, 1, PI_NET, "irdma_defer_t");
616 taskqueue_enqueue(hdl->deferred_tq, &hdl->deferred_task);
621 irdma_del_handler(hdl);
626 kfree(hdl);
641 struct irdma_handler *hdl;
647 hdl = irdma_find_handler(peer);
648 if (!hdl)
651 iwdev = hdl->iwdev;
658 taskqueue_enqueue(hdl->deferred_tq, &hdl->deferred_task);
660 taskqueue_drain(hdl->deferred_tq, &hdl->deferred_task);
661 taskqueue_free(hdl->deferred_tq);
662 hdl->iwdev->rf->dev_ctx.task_arg.iwdev = NULL;
663 hdl->iwdev->rf->dev_ctx.task_arg.peer = NULL;
666 hdl->iwdev->rf->tun_info.irdma_sysctl_tree = NULL;
667 hdl->iwdev->rf->tun_info.sws_sysctl_tree = NULL;
673 irdma_del_handler(iwdev->hdl);
674 kfree(iwdev->hdl);
737 struct irdma_handler *hdl;
744 list_for_each_entry(hdl, &irdma_handlers, list) {
745 if (!hdl->iwdev->rf->peer_info)
751 if (!hdl || !hdl_valid)
753 IRDMA_CLOSE(hdl->iwdev->rf->peer_info);
754 IRDMA_REMOVE(hdl->iwdev->rf->peer_info);