• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/infiniband/core/

Lines Matching defs:work

185 	struct work_struct	work;
193 struct work_struct work;
199 struct work_struct work;
1605 struct cma_work *work = context;
1608 route = &work->id->id.route;
1614 work->old_state = CMA_ROUTE_QUERY;
1615 work->new_state = CMA_ADDR_RESOLVED;
1616 work->event.event = RDMA_CM_EVENT_ROUTE_ERROR;
1617 work->event.status = status;
1620 queue_work(cma_wq, &work->work);
1624 struct cma_work *work)
1662 work, &id_priv->query);
1669 struct cma_work *work = container_of(_work, struct cma_work, work);
1670 struct rdma_id_private *id_priv = work->id;
1674 if (!cma_comp_exch(id_priv, work->old_state, work->new_state))
1677 if (id_priv->id.event_handler(&id_priv->id, &work->event)) {
1686 kfree(work);
1691 struct cma_ndev_work *work = container_of(_work, struct cma_ndev_work, work);
1692 struct rdma_id_private *id_priv = work->id;
1700 if (id_priv->id.event_handler(&id_priv->id, &work->event)) {
1710 kfree(work);
1716 struct cma_work *work;
1719 work = kzalloc(sizeof *work, GFP_KERNEL);
1720 if (!work)
1723 work->id = id_priv;
1724 INIT_WORK(&work->work, cma_work_handler);
1725 work->old_state = CMA_ROUTE_QUERY;
1726 work->new_state = CMA_ROUTE_RESOLVED;
1727 work->event.event = RDMA_CM_EVENT_ROUTE_RESOLVED;
1735 ret = cma_query_ib_route(id_priv, timeout_ms, work);
1744 kfree(work);
1774 struct cma_work *work;
1776 work = kzalloc(sizeof *work, GFP_KERNEL);
1777 if (!work)
1780 work->id = id_priv;
1781 INIT_WORK(&work->work, cma_work_handler);
1782 work->old_state = CMA_ROUTE_QUERY;
1783 work->new_state = CMA_ROUTE_RESOLVED;
1784 work->event.event = RDMA_CM_EVENT_ROUTE_RESOLVED;
1785 queue_work(cma_wq, &work->work);
1798 struct cma_work *work;
1808 work = kzalloc(sizeof *work, GFP_KERNEL);
1809 if (!work)
1812 work->id = id_priv;
1813 INIT_WORK(&work->work, cma_work_handler);
1856 work->old_state = CMA_ROUTE_QUERY;
1857 work->new_state = CMA_ROUTE_RESOLVED;
1858 work->event.event = RDMA_CM_EVENT_ROUTE_RESOLVED;
1859 work->event.status = 0;
1861 queue_work(cma_wq, &work->work);
1869 kfree(work);
2007 struct cma_work *work;
2012 work = kzalloc(sizeof *work, GFP_KERNEL);
2013 if (!work)
2037 work->id = id_priv;
2038 INIT_WORK(&work->work, cma_work_handler);
2039 work->old_state = CMA_ADDR_QUERY;
2040 work->new_state = CMA_ADDR_RESOLVED;
2041 work->event.event = RDMA_CM_EVENT_ADDR_RESOLVED;
2042 queue_work(cma_wq, &work->work);
2045 kfree(work);
3018 static void iboe_mcast_work_handler(struct work_struct *work)
3020 struct iboe_mcast_work *mw = container_of(work, struct iboe_mcast_work, work);
3059 struct iboe_mcast_work *work;
3068 work = kzalloc(sizeof *work, GFP_KERNEL);
3069 if (!work)
3104 work->id = id_priv;
3105 work->mc = mc;
3106 INIT_WORK(&work->work, iboe_mcast_work_handler);
3108 queue_work(cma_wq, &work->work);
3115 kfree(work);
3212 struct cma_ndev_work *work;
3227 work = kzalloc(sizeof *work, GFP_KERNEL);
3228 if (!work)
3231 INIT_WORK(&work->work, cma_ndev_work_handler);
3232 work->id = id_priv;
3233 work->event.event = RDMA_CM_EVENT_ADDR_CHANGE;
3235 queue_work(cma_wq, &work->work);