Lines Matching refs:work

338 	struct work_struct	work;
346 struct work_struct work;
352 struct work_struct work;
2349 struct cma_work *work = context;
2352 route = &work->id->id.route;
2358 work->old_state = RDMA_CM_ROUTE_QUERY;
2359 work->new_state = RDMA_CM_ADDR_RESOLVED;
2360 work->event.event = RDMA_CM_EVENT_ROUTE_ERROR;
2361 work->event.status = status;
2364 queue_work(cma_wq, &work->work);
2368 struct cma_work *work)
2409 work, &id_priv->query);
2416 struct cma_work *work = container_of(_work, struct cma_work, work);
2417 struct rdma_id_private *id_priv = work->id;
2421 if (!cma_comp_exch(id_priv, work->old_state, work->new_state))
2424 if (id_priv->id.event_handler(&id_priv->id, &work->event)) {
2433 kfree(work);
2439 struct cma_work *work;
2442 work = kzalloc(sizeof *work, GFP_KERNEL);
2443 if (!work)
2446 work->id = id_priv;
2447 INIT_WORK(&work->work, cma_work_handler);
2448 work->old_state = RDMA_CM_ROUTE_QUERY;
2449 work->new_state = RDMA_CM_ROUTE_RESOLVED;
2450 work->event.event = RDMA_CM_EVENT_ROUTE_RESOLVED;
2458 ret = cma_query_ib_route(id_priv, timeout_ms, work);
2467 kfree(work);
2499 struct cma_work *work;
2501 work = kzalloc(sizeof *work, GFP_KERNEL);
2502 if (!work)
2505 work->id = id_priv;
2506 INIT_WORK(&work->work, cma_work_handler);
2507 work->old_state = RDMA_CM_ROUTE_QUERY;
2508 work->new_state = RDMA_CM_ROUTE_RESOLVED;
2509 work->event.event = RDMA_CM_EVENT_ROUTE_RESOLVED;
2510 queue_work(cma_wq, &work->work);
2539 struct cma_work *work;
2544 work = kzalloc(sizeof *work, GFP_KERNEL);
2545 if (!work)
2548 work->id = id_priv;
2549 INIT_WORK(&work->work, cma_work_handler);
2614 work->old_state = RDMA_CM_ROUTE_QUERY;
2615 work->new_state = RDMA_CM_ROUTE_RESOLVED;
2616 work->event.event = RDMA_CM_EVENT_ROUTE_RESOLVED;
2617 work->event.status = 0;
2619 queue_work(cma_wq, &work->work);
2627 kfree(work);
2775 struct cma_work *work;
2779 work = kzalloc(sizeof *work, GFP_KERNEL);
2780 if (!work)
2792 work->id = id_priv;
2793 INIT_WORK(&work->work, cma_work_handler);
2794 work->old_state = RDMA_CM_ADDR_QUERY;
2795 work->new_state = RDMA_CM_ADDR_RESOLVED;
2796 work->event.event = RDMA_CM_EVENT_ADDR_RESOLVED;
2797 queue_work(cma_wq, &work->work);
2800 kfree(work);
2806 struct cma_work *work;
2809 work = kzalloc(sizeof *work, GFP_KERNEL);
2810 if (!work)
2822 work->id = id_priv;
2823 INIT_WORK(&work->work, cma_work_handler);
2824 work->old_state = RDMA_CM_ADDR_QUERY;
2825 work->new_state = RDMA_CM_ADDR_RESOLVED;
2826 work->event.event = RDMA_CM_EVENT_ADDR_RESOLVED;
2827 queue_work(cma_wq, &work->work);
2830 kfree(work);
4001 static void iboe_mcast_work_handler(struct work_struct *work)
4003 struct iboe_mcast_work *mw = container_of(work, struct iboe_mcast_work, work);
4045 struct iboe_mcast_work *work;
4058 work = kzalloc(sizeof *work, GFP_KERNEL);
4059 if (!work)
4108 work->id = id_priv;
4109 work->mc = mc;
4110 INIT_WORK(&work->work, iboe_mcast_work_handler);
4112 queue_work(cma_wq, &work->work);
4119 kfree(work);