Lines Matching defs:inbox

586 			      struct mlx4_cmd_mailbox *inbox)
588 u8 sched = *(u8 *)(inbox->buf + 64);
589 u8 orig_index = *(u8 *)(inbox->buf + 35);
597 *(u8 *)(inbox->buf + 35) = new_index;
600 static void update_gid(struct mlx4_dev *dev, struct mlx4_cmd_mailbox *inbox,
603 struct mlx4_qp_context *qp_ctx = inbox->buf + 8;
604 enum mlx4_qp_optpar optpar = be32_to_cpu(*(__be32 *) inbox->buf);
662 struct mlx4_cmd_mailbox *inbox,
665 struct mlx4_qp_context *qpc = inbox->buf + 8;
2020 struct mlx4_cmd_mailbox *inbox,
2318 struct mlx4_cmd_mailbox *inbox,
2465 struct mlx4_cmd_mailbox *inbox,
2473 int mtt_base = mr_get_mtt_addr(inbox->buf) / dev->caps.mtt_entry_sz;
2487 if (!mr_is_region(inbox->buf)) {
2493 pd = mr_get_pd(inbox->buf);
2500 if (mr_is_fmr(inbox->buf)) {
2502 if (mr_is_bind_enabled(inbox->buf)) {
2507 if (!mr_is_region(inbox->buf)) {
2513 phys = mr_phys_mpt(inbox->buf);
2520 mr_get_mtt_size(inbox->buf), mtt);
2527 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2550 struct mlx4_cmd_mailbox *inbox,
2564 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2582 struct mlx4_cmd_mailbox *inbox,
2601 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2638 struct mlx4_cmd_mailbox *inbox,
2646 struct mlx4_qp_context *qpc = inbox->buf + 8;
2697 update_pkey_index(dev, slave, inbox);
2698 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2772 struct mlx4_cmd_mailbox *inbox,
2779 struct mlx4_eq_context *eqc = inbox->buf;
2800 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2844 struct mlx4_cmd_mailbox *inbox,
2853 qp_ctx = inbox->buf + 8;
2855 optpar = be32_to_cpu(*(__be32 *) inbox->buf);
2900 struct mlx4_cmd_mailbox *inbox,
2905 __be64 *page_list = inbox->buf;
2919 * - Translate inbox contents to simple addresses in host endianess */
2938 struct mlx4_cmd_mailbox *inbox,
2955 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3044 struct mlx4_cmd_mailbox *inbox,
3062 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3071 struct mlx4_cmd_mailbox *inbox,
3077 struct mlx4_cq_context *cqc = inbox->buf;
3091 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3109 struct mlx4_cmd_mailbox *inbox,
3120 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3134 struct mlx4_cmd_mailbox *inbox,
3149 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3158 struct mlx4_cmd_mailbox *inbox,
3166 struct mlx4_cq_context *cqc = inbox->buf;
3185 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3206 struct mlx4_cmd_mailbox *inbox,
3222 err = handle_resize(dev, slave, vhcr, inbox, outbox, cmd, cq);
3226 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3247 struct mlx4_cmd_mailbox *inbox,
3255 struct mlx4_srq_context *srqc = inbox->buf;
3272 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3292 struct mlx4_cmd_mailbox *inbox,
3303 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3321 struct mlx4_cmd_mailbox *inbox,
3336 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3344 struct mlx4_cmd_mailbox *inbox,
3361 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3369 struct mlx4_cmd_mailbox *inbox,
3385 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3393 struct mlx4_cmd_mailbox *inbox,
3397 struct mlx4_qp_context *context = inbox->buf + 8;
3399 update_pkey_index(dev, slave, inbox);
3400 return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3405 struct mlx4_cmd_mailbox *inbox)
3410 u8 sched = *(u8 *)(inbox->buf + 64);
3424 struct mlx4_cmd_mailbox *inbox,
3429 struct mlx4_qp_context *qpc = inbox->buf + 8;
3440 err = verify_qp_parameters(dev, inbox, QP_TRANS_INIT2RTR, slave);
3444 if (roce_verify_mac(dev, slave, qpc, inbox))
3447 update_pkey_index(dev, slave, inbox);
3448 update_gid(dev, inbox, (u8)slave);
3462 err = update_vport_qp_param(dev, inbox, slave, qpn);
3467 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3488 struct mlx4_cmd_mailbox *inbox,
3493 struct mlx4_qp_context *context = inbox->buf + 8;
3495 err = verify_qp_parameters(dev, inbox, QP_TRANS_RTR2RTS, slave);
3499 update_pkey_index(dev, slave, inbox);
3500 update_gid(dev, inbox, (u8)slave);
3502 return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3507 struct mlx4_cmd_mailbox *inbox,
3512 struct mlx4_qp_context *context = inbox->buf + 8;
3514 err = verify_qp_parameters(dev, inbox, QP_TRANS_RTS2RTS, slave);
3518 update_pkey_index(dev, slave, inbox);
3519 update_gid(dev, inbox, (u8)slave);
3521 return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3527 struct mlx4_cmd_mailbox *inbox,
3531 struct mlx4_qp_context *context = inbox->buf + 8;
3533 return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3538 struct mlx4_cmd_mailbox *inbox,
3543 struct mlx4_qp_context *context = inbox->buf + 8;
3545 err = verify_qp_parameters(dev, inbox, QP_TRANS_SQD2SQD, slave);
3550 update_gid(dev, inbox, (u8)slave);
3551 update_pkey_index(dev, slave, inbox);
3552 return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3557 struct mlx4_cmd_mailbox *inbox,
3562 struct mlx4_qp_context *context = inbox->buf + 8;
3564 err = verify_qp_parameters(dev, inbox, QP_TRANS_SQD2RTS, slave);
3569 update_gid(dev, inbox, (u8)slave);
3570 update_pkey_index(dev, slave, inbox);
3571 return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3576 struct mlx4_cmd_mailbox *inbox,
3587 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
3700 struct mlx4_cmd_mailbox *inbox,
3705 u8 *gid = inbox->buf;
3782 struct mlx4_cmd_mailbox *inbox,
3794 ctrl = (struct mlx4_net_trans_rule_hw_ctrl *)inbox->buf;
3798 /* Clear a space in the inbox for eth header */
3839 struct mlx4_cmd_mailbox *inbox,
3858 ctrl = (struct mlx4_net_trans_rule_hw_ctrl *)inbox->buf;
3881 if (add_eth_header(dev, slave, inbox, rlist, header_id)) {
3894 err = mlx4_cmd_imm(dev, inbox->dma, &vhcr->out_param,
3918 struct mlx4_cmd_mailbox *inbox,
3964 struct mlx4_cmd_mailbox *inbox,
3970 err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);