/linux-master/net/core/ |
H A D | netdev-genl.c | 293 u32 q_idx, u32 q_type, const struct genl_info *info) 304 nla_put_u32(rsp, NETDEV_A_QUEUE_TYPE, q_type) || 308 switch (q_type) { 332 u32 q_type) 334 switch (q_type) { 348 u32 q_type, const struct genl_info *info) 355 err = netdev_nl_queue_validate(netdev, q_idx, q_type); 359 return netdev_nl_queue_fill_one(rsp, netdev, q_idx, q_type, info); 364 u32 q_id, q_type, ifindex; local 375 q_type 292 netdev_nl_queue_fill_one(struct sk_buff *rsp, struct net_device *netdev, u32 q_idx, u32 q_type, const struct genl_info *info) argument 331 netdev_nl_queue_validate(struct net_device *netdev, u32 q_id, u32 q_type) argument 347 netdev_nl_queue_fill(struct sk_buff *rsp, struct net_device *netdev, u32 q_idx, u32 q_type, const struct genl_info *info) argument 507 netdev_nl_stats_queue(struct net_device *netdev, struct sk_buff *rsp, u32 q_type, int i, const struct genl_info *info) argument [all...] |
/linux-master/fs/xfs/scrub/ |
H A D | quotacheck.c | 43 xfs_dqtype_t q_type; member in struct:xqcheck_dqtrx 187 xfs_dqtype_t q_type, 193 if (dqa->dqtrx[i].q_type == 0 || 194 (dqa->dqtrx[i].q_type == q_type && 233 switch (p->q_type) { 271 dqtrx = xqcheck_get_dqtrx(dqa, p->q_type, p->q_id); 274 if (dqtrx->q_type == 0) { 275 dqtrx->q_type = p->q_type; 185 xqcheck_get_dqtrx( struct xqcheck_dqacct *dqa, xfs_dqtype_t q_type, xfs_dqid_t q_id) argument [all...] |
H A D | quotacheck_repair.c | 111 trace_xrep_quotacheck_dquot(xqc->sc->mp, dq->q_type, dq->q_id);
|
H A D | quota_repair.c | 85 trace_xrep_dquot_item_fill_bmap_hole(sc->mp, dq->q_type, dq->q_id); 94 xfs_qm_init_dquot_blk(sc->tp, dq->q_id, dq->q_type, bp); 254 trace_xrep_dquot_item(sc->mp, dq->q_type, dq->q_id);
|
/linux-master/fs/xfs/ |
H A D | xfs_dquot.h | 65 xfs_dqtype_t q_type; member in struct:xfs_dquot 136 return dqp->q_type & XFS_DQTYPE_REC_MASK;
|
H A D | xfs_qm.h | 118 xfs_dqtype_t q_type; member in struct:xfs_mod_ino_dqtrx_params
|
H A D | xfs_trans_dquot.c | 63 !(dqp->q_type & XFS_DQTYPE_BIGTIME)) 64 dqp->q_type |= XFS_DQTYPE_BIGTIME; 165 .q_type = xfs_dquot_type(dqp), 445 .q_type = xfs_dquot_type(dqp), 614 .q_type = xfs_dquot_type(dqp),
|
H A D | xfs_dquot.c | 55 switch (dqp->q_type) { 503 dqp->q_type = type; 608 dqp->q_type = ddqp->d_type; 645 ddqp->d_type = dqp->q_type; 1229 if (dqp->q_type & XFS_DQTYPE_BIGTIME) {
|
H A D | xfs_quota.h | 89 xfs_dqtype_t q_type; member in struct:xfs_apply_dqtrx_params
|
/linux-master/drivers/crypto/ |
H A D | n2_core.c | 58 u8 q_type; member in struct:spu_queue 82 if (q->q_type == HV_NCS_QTYPE_MAU) { 1539 static void *new_queue(unsigned long q_type) argument 1541 return kmem_cache_zalloc(queue_cache[q_type - 1], GFP_KERNEL); 1544 static void free_queue(void *p, unsigned long q_type) argument 1546 kmem_cache_free(queue_cache[q_type - 1], p); 1586 unsigned long q_type = qr->type; local 1589 hv_ret = sun4v_ncs_qconf(q_type, __pa(p->q), 1597 static int spu_queue_register(struct spu_queue *p, unsigned long q_type) argument 1600 struct spu_qreg qr = { .queue = p, .type = q_type }; 1687 handle_exec_unit(struct spu_mdesc_info *ip, struct list_head *list, struct platform_device *dev, struct mdesc_handle *mdesc, u64 node, const char *iname, unsigned long q_type, irq_handler_t handler, struct spu_queue **table) argument 1719 spu_mdesc_scan(struct mdesc_handle *mdesc, struct platform_device *dev, struct spu_mdesc_info *ip, struct list_head *list, const char *exec_name, unsigned long q_type, irq_handler_t handler, struct spu_queue **table) argument [all...] |
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_virtchnl.c | 1048 * @q_type: queue model 1056 static int idpf_vport_get_q_reg(u32 *reg_vals, int num_regs, u32 q_type, argument 1068 if (le32_to_cpu(chunk->type) != q_type) 1087 * @q_type: queue model 1092 int num_regs, u32 q_type) 1098 switch (q_type) { 1452 cpu_to_le32(tx_qgrp->txqs[j]->q_type); 1481 qi[k].type = cpu_to_le32(tx_qgrp->complq->q_type); 1575 qi[k].type = cpu_to_le32(bufq->q_type); 1622 qi[k].type = cpu_to_le32(rxq->q_type); 1091 __idpf_queue_reg_init(struct idpf_vport *vport, u32 *reg_vals, int num_regs, u32 q_type) argument 3199 idpf_vport_get_queue_ids(u32 *qids, int num_qids, u16 q_type, struct virtchnl2_queue_reg_chunks *chunks) argument 3240 __idpf_vport_queue_ids_init(struct idpf_vport *vport, const u32 *qids, int num_qids, u32 q_type) argument 3322 u16 q_type; local [all...] |
H A D | idpf_ethtool.c | 699 if (q->q_type == VIRTCHNL2_QUEUE_TYPE_RX) { 1028 if (q->q_type == VIRTCHNL2_QUEUE_TYPE_RX) {
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_cmdq.h | 117 u32 q_type; member in struct:hinic_cmdq_arm_bit
|
H A D | hinic_hw_cmdq.c | 502 * @q_type: type of queue to set the arm bit for 508 enum hinic_set_arm_qtype q_type, u32 q_id) 516 arm_bit.q_type = q_type; 733 * @q_type: the cmdq type of the cmdq 739 enum hinic_cmdq_type q_type, void __iomem *db_area) 744 cmdq->cmdq_type = q_type; 507 hinic_set_arm_bit(struct hinic_cmdqs *cmdqs, enum hinic_set_arm_qtype q_type, u32 q_id) argument 738 init_cmdq(struct hinic_cmdq *cmdq, struct hinic_wq *wq, enum hinic_cmdq_type q_type, void __iomem *db_area) argument
|
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | hw_channel.c | 483 enum gdma_queue_type q_type, u16 q_depth, 492 WARN_ON(q_type != GDMA_SQ && q_type != GDMA_RQ); 494 if (q_type == GDMA_RQ) 506 err = mana_hwc_create_gdma_wq(hwc, q_type, queue_size, &queue); 482 mana_hwc_create_wq(struct hw_channel_context *hwc, enum gdma_queue_type q_type, u16 q_depth, u32 max_msg_size, struct hwc_cq *hwc_cq, struct hwc_wq **hwc_wq_ptr) argument
|
H A D | gdma_main.c | 274 enum gdma_queue_type q_type, u32 qid, 280 switch (q_type) { 1026 enum gdma_queue_type q_type, 1048 if (q_type == GDMA_SQ) 273 mana_gd_ring_doorbell(struct gdma_context *gc, u32 db_index, enum gdma_queue_type q_type, u32 qid, u32 tail_ptr, u8 num_req) argument 1025 mana_gd_write_client_oob(const struct gdma_wqe_request *wqe_req, enum gdma_queue_type q_type, u32 client_oob_size, u32 sgl_data_size, u8 *wqe_ptr) argument
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 599 * @q_type: specific Control queue type 610 static int ice_init_ctrlq(struct ice_hw *hw, enum ice_ctl_q q_type) argument 615 switch (q_type) { 631 cq->qtype = q_type; 686 * @q_type: specific Control queue type 690 static void ice_shutdown_ctrlq(struct ice_hw *hw, enum ice_ctl_q q_type) argument 694 switch (q_type) {
|
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_dquot_buf.c | 321 if (timer != 0 && (dqp->q_type & XFS_DQTYPE_BIGTIME))
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | puda.h | 31 u8 q_type; member in struct:irdma_puda_cmpl_info
|
H A D | user.h | 248 u8 q_type; member in struct:irdma_cq_poll_info
|
H A D | puda.c | 277 info->q_type = (u8)FIELD_GET(IRDMA_CQ_SQ, qword3); 286 if (info->q_type == IRDMA_CQE_QTYPE_RQ) { 361 if (info.q_type == IRDMA_CQE_QTYPE_RQ) {
|
/linux-master/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpni.h | 952 enum dpni_queue_type q_type, 961 enum dpni_queue_type q_type,
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci.c | 185 enum mlxsw_pci_queue_type q_type) 187 return &mlxsw_pci->queues[q_type]; 191 enum mlxsw_pci_queue_type q_type) 195 queue_group = mlxsw_pci_queue_type_group_get(mlxsw_pci, q_type); 211 enum mlxsw_pci_queue_type q_type, u8 q_num) 213 return &mlxsw_pci->queues[q_type].q[q_num]; 184 mlxsw_pci_queue_type_group_get(struct mlxsw_pci *mlxsw_pci, enum mlxsw_pci_queue_type q_type) argument 190 __mlxsw_pci_queue_count(struct mlxsw_pci *mlxsw_pci, enum mlxsw_pci_queue_type q_type) argument 210 __mlxsw_pci_queue_get(struct mlxsw_pci *mlxsw_pci, enum mlxsw_pci_queue_type q_type, u8 q_num) argument
|
/linux-master/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_sriov.c | 170 unsigned long q_type) 216 if (test_bit(BNX2X_Q_TYPE_HAS_RX, &q_type)) { 228 if (test_bit(BNX2X_Q_TYPE_HAS_TX, &q_type)) { 1419 unsigned long q_type = 0; local 1421 set_bit(BNX2X_Q_TYPE_HAS_TX, &q_type); 1422 set_bit(BNX2X_Q_TYPE_HAS_RX, &q_type); 1429 q_type); 166 bnx2x_vfop_qctor_prep(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_queue *q, struct bnx2x_vf_queue_construct_params *p, unsigned long q_type) argument
|
/linux-master/drivers/accel/habanalabs/common/ |
H A D | command_submission.c | 1940 enum hl_queue_type q_type, u32 q_idx, u32 encaps_signal_offset) 1949 job = hl_cs_allocate_job(hdev, q_type, true); 1962 cb = hl_cb_kernel_create(hdev, cb_size, q_type == QUEUE_TYPE_HW); 2216 enum hl_queue_type q_type; local 2243 q_type = hw_queue_prop->type; 2406 rc = cs_ioctl_signal_wait_create_jobs(hdev, ctx, cs, q_type, 2421 if (q_type == QUEUE_TYPE_HW) 1938 cs_ioctl_signal_wait_create_jobs(struct hl_device *hdev, struct hl_ctx *ctx, struct hl_cs *cs, enum hl_queue_type q_type, u32 q_idx, u32 encaps_signal_offset) argument
|