Searched refs:eqe (Results 1 - 25 of 56) sorted by relevance

123

/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Deq.c123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); local
124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe;
129 struct mlx4_eqe *eqe = local
131 return (!!(eqe->owner & 0x80) ^
133 eqe : NULL;
146 struct mlx4_eqe *eqe; local
150 for (eqe = next_slave_event_eqe(slave_eq); eqe;
151 eqe
194 slave_event(struct mlx4_dev *dev, u8 slave, struct mlx4_eqe *eqe) argument
223 mlx4_slave_event(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe) argument
255 struct mlx4_eqe eqe; local
275 struct mlx4_eqe eqe; local
293 struct mlx4_eqe eqe; local
434 struct mlx4_eqe eqe; local
497 struct mlx4_eqe *eqe; local
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Devents.c143 /* handles all FW events, type == eqe->type */
149 struct mlx5_eqe *eqe = data; local
151 mlx5_core_dbg(events->dev, "Async eqe type %s, subtype (%d)\n",
152 eqe_type_str(eqe->type), eqe->sub_type);
161 struct mlx5_eqe *eqe = data; local
165 value_lsb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_lsb);
166 value_msb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_msb);
223 struct mlx5_eqe *eqe = data; local
231 module_event_eqe = &eqe
314 struct mlx5_eqe *eqe = data; local
337 struct mlx5_eqe *eqe = data; local
[all...]
H A Deq.c115 struct mlx5_eqe *eqe; local
119 eqe = next_eqe_sw(eq);
120 if (!eqe)
130 /* Assume (eqe->type) is always MLX5_EVENT_TYPE_COMP */
131 cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff;
136 cq->comp(cq, eqe);
145 } while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq)));
207 struct mlx5_eqe *eqe; local
218 eqe = next_eqe_sw(eq);
219 if (!eqe)
255 struct mlx5_eqe *eqe; local
506 struct mlx5_eqe *eqe; local
790 struct mlx5_eqe *eqe; local
[all...]
H A Dfw_reset.c631 static void mlx5_sync_reset_events_handle(struct mlx5_fw_reset *fw_reset, struct mlx5_eqe *eqe) argument
636 sync_fw_update_eqe = &eqe->data.sync_fw_update;
657 struct mlx5_eqe *eqe = data; local
662 switch (eqe->sub_type) {
667 mlx5_sync_reset_events_handle(fw_reset, eqe);
H A Dpagealloc.c617 struct mlx5_eqe *eqe; local
625 eqe = data;
627 func_id = be16_to_cpu(eqe->data.req_pages.func_id);
628 npages = be32_to_cpu(eqe->data.req_pages.num_pages);
629 ec_function = be16_to_cpu(eqe->data.req_pages.ec_function) & EC_FUNCTION_MASK;
630 release_all = be16_to_cpu(eqe->data.req_pages.ec_function) &
H A Dcq.c70 struct mlx5_eqe *eqe)
69 mlx5_add_cq_to_tasklet(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe) argument
H A Den_txrx.c258 void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) argument
/linux-master/drivers/net/ethernet/mellanox/mlxsw/
H A Dpci_hw.h377 MLXSW_ITEM32(pci, eqe, event_type, 0x0C, 24, 8);
384 MLXSW_ITEM32(pci, eqe, event_sub_type, 0x0C, 16, 8);
389 MLXSW_ITEM32(pci, eqe, cqn, 0x0C, 8, 7);
394 MLXSW_ITEM32(pci, eqe, owner, 0x0C, 0, 1);
399 MLXSW_ITEM32(pci, eqe, cmd_token, 0x00, 16, 16);
404 MLXSW_ITEM32(pci, eqe, cmd_status, 0x00, 0, 8);
409 MLXSW_ITEM32(pci, eqe, cmd_out_param_h, 0x04, 0, 32);
414 MLXSW_ITEM32(pci, eqe, cmd_out_param_l, 0x08, 0, 32);
/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_eq.c236 struct mthca_eqe *eqe; local
237 eqe = get_eqe(eq, eq->cons_index);
238 return (MTHCA_EQ_ENTRY_OWNER_HW & eqe->owner) ? NULL : eqe;
241 static inline void set_eqe_hw(struct mthca_eqe *eqe) argument
243 eqe->owner = MTHCA_EQ_ENTRY_OWNER_HW;
262 struct mthca_eqe *eqe; local
267 while ((eqe = next_eqe_sw(eq))) {
274 switch (eqe->type) {
276 disarm_cqn = be32_to_cpu(eqe
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lib/
H A Deq.h63 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & eq->fbc.sz_m1); local
65 return (eqe->owner ^ (eq->cons_index >> eq->fbc.log_sz)) & 1 ? NULL : eqe;
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dcore.c163 static int fpga_err_event(struct notifier_block *nb, unsigned long event, void *eqe) argument
167 return mlx5_fpga_event(fdev, event, eqe);
170 static int fpga_qp_err_event(struct notifier_block *nb, unsigned long event, void *eqe) argument
174 return mlx5_fpga_event(fdev, event, eqe);
334 unsigned long event, void *eqe)
336 void *data = ((struct mlx5_eqe *)eqe)->data.raw;
333 mlx5_fpga_event(struct mlx5_fpga_device *fdev, unsigned long event, void *eqe) argument
/linux-master/include/linux/mlx5/
H A Dcq.h49 void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
57 void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
/linux-master/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_main.c413 struct pvrdma_eqe *eqe; local
415 eqe = get_eqe(dev, head);
417 switch (eqe->type) {
426 pvrdma_qp_event(dev, eqe->info, eqe->type);
430 pvrdma_cq_event(dev, eqe->info, eqe->type);
435 pvrdma_srq_event(dev, eqe->info, eqe->type);
445 pvrdma_dev_event(dev, eqe
[all...]
/linux-master/drivers/infiniband/hw/efa/
H A Defa_main.c75 static void efa_process_comp_eqe(struct efa_dev *dev, struct efa_admin_eqe *eqe) argument
77 u16 cqn = eqe->u.comp_event.cqn;
92 static void efa_process_eqe(struct efa_com_eq *eeq, struct efa_admin_eqe *eqe) argument
96 if (likely(EFA_GET(&eqe->common, EFA_ADMIN_EQE_EVENT_TYPE) ==
98 efa_process_comp_eqe(dev, eqe);
102 EFA_GET(&eqe->common,
H A Defa_com.h85 struct efa_admin_eqe *eqe);
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_eq.c24 u64 *eqe = get_queue_entry(eq->qbuf, eq->ci, eq->depth, EQE_SHIFT); local
25 u32 owner = FIELD_GET(ERDMA_CEQE_HDR_O_MASK, READ_ONCE(*eqe));
27 return owner ^ !!(eq->ci & eq->depth) ? eqe : NULL;
/linux-master/drivers/infiniband/hw/mlx4/
H A Dmad.c63 #define GET_BLK_PTR_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.block_ptr)
64 #define GET_MASK_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.tbl_entries_mask)
1120 struct mlx4_eqe *eqe)
1122 __propagate_pkey_ev(dev, port_num, GET_BLK_PTR_FROM_EQE(eqe),
1123 GET_MASK_FROM_EQE(eqe));
1181 struct mlx4_eqe *eqe = &(ew->ib_eqe); local
1182 u32 port = eqe->event.port_mgmt_change.port;
1187 switch (eqe
1119 propagate_pkey_ev(struct mlx4_ib_dev *dev, int port_num, struct mlx4_eqe *eqe) argument
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dmonitor_stats.c72 unsigned long event, void *eqe)
71 mlx5e_monitor_event_handler(struct notifier_block *nb, unsigned long event, void *eqe) argument
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/sf/
H A Dvhca_event.c122 struct mlx5_eqe *eqe = data; local
130 work->event.function_id = be16_to_cpu(eqe->data.vhca_state.function_id);
/linux-master/drivers/infiniband/hw/mlx5/
H A Dqpc.c91 static int dct_event_notifier(struct mlx5_ib_dev *dev, struct mlx5_eqe *eqe) argument
97 qpn = be32_to_cpu(eqe->data.dct.dctn) & 0xFFFFFF;
112 struct mlx5_eqe *eqe = data; local
119 return dct_event_notifier(dev, eqe);
128 rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
129 rsn |= (eqe->data.qp_srq.type << MLX5_USER_INDEX_LEN);
H A Ddevx.c251 static u16 get_event_obj_type(unsigned long event_type, struct mlx5_eqe *eqe) argument
264 return eqe->data.qp_srq.type;
272 return MLX5_GET(affiliated_event_header, &eqe->data, obj_type);
1429 static void devx_cq_comp(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) argument
1447 dispatch_event_fd(&obj_event->obj_sub_list, eqe);
2414 struct mlx5_eqe *eqe = data; local
2428 obj_id = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
2431 obj_id = be32_to_cpu(eqe->data.xrq_err.type_xrqn) & 0xffffff;
2435 obj_id = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff;
2438 obj_id = be32_to_cpu(eqe
[all...]
H A Dsrq_cmd.c728 struct mlx5_eqe *eqe; local
737 eqe = data;
738 srqn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
749 srq->event(srq, eqe->type);
/linux-master/drivers/net/ethernet/ibm/ehea/
H A Dehea_main.c925 struct ehea_eqe *eqe; local
931 eqe = ehea_poll_eq(port->qp_eq);
933 while (eqe) {
934 qp_token = EHEA_BMASK_GET(EHEA_EQE_QP_TOKEN, eqe->entry);
936 eqe->entry, qp_token);
950 eqe = ehea_poll_eq(port->qp_eq);
1138 static void ehea_parse_eqe(struct ehea_adapter *adapter, u64 eqe) argument
1146 ec = EHEA_BMASK_GET(NEQE_EVENT_CODE, eqe);
1147 portnum = EHEA_BMASK_GET(NEQE_PORTNUM, eqe);
1158 if (EHEA_BMASK_GET(NEQE_PORT_UP, eqe)) {
1222 struct ehea_eqe *eqe; local
[all...]
/linux-master/drivers/net/ethernet/microsoft/mana/
H A Dgdma_main.c353 struct gdma_eqe *eqe; local
356 eqe = &eq_eqe_ptr[head];
357 eqe_info.as_uint32 = eqe->eqe_info;
362 cq_id = eqe->details[0] & 0xFFFFFF;
387 memcpy(&event.details, &eqe->details, GDMA_EVENT_DATA_SIZE);
403 struct gdma_eqe *eqe; local
414 eqe = &eq_eqe_ptr[eq->head % num_eqe];
415 eqe_info.as_uint32 = eqe->eqe_info;
434 * reading eqe.
/linux-master/drivers/scsi/be2iscsi/
H A Dbe_main.c686 struct be_eq_entry *eqe; local
696 eqe = queue_tail_node(eq);
699 while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
701 if (((eqe->dw[offsetof(struct amap_eq_entry,
706 AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
708 eqe = queue_tail_node(eq);
750 struct be_eq_entry *eqe; local
771 eqe = queue_tail_node(eq);
775 while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
777 if (((eqe
2053 struct be_eq_entry *eqe = NULL; local
3600 struct be_eq_entry *eqe = NULL; local
[all...]

Completed in 266 milliseconds

123