Lines Matching refs:shampo

313 	rq->mpwqe.shampo = kvzalloc_node(sizeof(*rq->mpwqe.shampo),
315 if (!rq->mpwqe.shampo)
322 kvfree(rq->mpwqe.shampo);
327 struct mlx5e_shampo_hd *shampo = rq->mpwqe.shampo;
329 shampo->bitmap = bitmap_zalloc_node(shampo->hd_per_wq, GFP_KERNEL,
331 shampo->info = kvzalloc_node(array_size(shampo->hd_per_wq,
332 sizeof(*shampo->info)),
334 shampo->pages = kvzalloc_node(array_size(shampo->hd_per_wq,
335 sizeof(*shampo->pages)),
337 if (!shampo->bitmap || !shampo->info || !shampo->pages)
343 kvfree(shampo->info);
344 kvfree(shampo->bitmap);
345 kvfree(shampo->pages);
352 kvfree(rq->mpwqe.shampo->bitmap);
353 kvfree(rq->mpwqe.shampo->info);
354 kvfree(rq->mpwqe.shampo->pages);
570 if (max_klm_size < rq->mpwqe.shampo->hd_per_wq) {
571 mlx5_core_err(mdev, "max klm list size 0x%x is smaller than shampo header buffer list size 0x%x\n",
572 max_klm_size, rq->mpwqe.shampo->hd_per_wq);
575 return mlx5e_create_umr_klm_mkey(mdev, rq->mpwqe.shampo->hd_per_wq,
576 &rq->mpwqe.shampo->mkey);
754 rq->mpwqe.shampo->hd_per_wq =
767 rq->mpwqe.shampo->key =
768 cpu_to_be32(rq->mpwqe.shampo->mkey);
769 rq->mpwqe.shampo->hd_per_wqe =
772 *pool_size += (rq->mpwqe.shampo->hd_per_wqe * wq_size) /
779 mlx5_core_destroy_mkey(mdev, rq->mpwqe.shampo->mkey);
793 mlx5_core_destroy_mkey(rq->mdev, rq->mpwqe.shampo->mkey);
1054 order_base_2(rq->mpwqe.shampo->hd_per_wq));
1055 MLX5_SET(wq, wq, headers_mkey, rq->mpwqe.shampo->mkey);
1214 len = (rq->mpwqe.shampo->pi - rq->mpwqe.shampo->ci) &
1215 (rq->mpwqe.shampo->hd_per_wq - 1);
1216 mlx5e_shampo_dealloc_hd(rq, len, rq->mpwqe.shampo->ci, false);
1217 rq->mpwqe.shampo->pi = rq->mpwqe.shampo->ci;
1247 mlx5e_shampo_dealloc_hd(rq, rq->mpwqe.shampo->hd_per_wq,
3996 new_params.packet_merge.shampo.match_criteria_type =
3998 new_params.packet_merge.shampo.alignment_granularity =