/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | tx.h | 11 int mlx5e_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags);
|
H A D | pool.h | 25 int mlx5e_xsk_setup_pool(struct net_device *dev, struct xsk_buff_pool *pool, u16 qid);
|
/linux-master/arch/s390/include/asm/ |
H A D | ap.h | 77 " lgr 0,%[reg0]\n" /* qid into gr0 */ 130 * @qid: The AP queue number 135 static inline struct ap_queue_status ap_tapq(ap_qid_t qid, 142 " lgr 0,%[qid]\n" /* qid into gr0 */ 148 : [qid] "d" (qid) 157 * @qid: The AP queue number 163 static inline struct ap_queue_status ap_test_queue(ap_qid_t qid, int tbit, argument 167 qid | 178 ap_rapq(ap_qid_t qid, int fbit) argument 203 ap_zapq(ap_qid_t qid, int fbit) argument 298 ap_aqic(ap_qid_t qid, union ap_qirq_ctrl qirqctrl, phys_addr_t pa_ind) argument 347 ap_qact(ap_qid_t qid, int ifbit, union ap_qact_ap_info *apinfo) argument 378 ap_bapq(ap_qid_t qid) argument 404 ap_aapq(ap_qid_t qid, unsigned int sec_idx) argument 434 ap_nqap(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length) argument 490 ap_dqap(ap_qid_t qid, unsigned long *psmid, void *msg, size_t msglen, size_t *length, size_t *reslength, unsigned long *resgr0) argument [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | qos.h | 24 void mlx5e_deactivate_qos_sq(struct mlx5e_priv *priv, u16 qid); 25 void mlx5e_close_qos_sq(struct mlx5e_priv *priv, u16 qid); 26 void mlx5e_reactivate_qos_sq(struct mlx5e_priv *priv, u16 qid, struct netdev_queue *txq); 27 void mlx5e_reset_qdisc(struct net_device *dev, u16 qid); 38 u16 mlx5e_qid_from_qos(struct mlx5e_channels *chs, u16 qid);
|
H A D | htb.c | 17 u16 qid; member in struct:mlx5e_qos_node 40 if (node->qid == MLX5E_QOS_QID_INNER) 42 err = callback(data, node->qid, node->hw_id); 70 mlx5e_htb_node_create_leaf(struct mlx5e_htb *htb, u16 classid, u16 qid, argument 81 node->qid = qid; 82 __set_bit(qid, htb->qos_used_qids); 100 node->qid = MLX5E_QOS_QID_INNER; 134 if (node->qid != MLX5E_QOS_QID_INNER) { 135 __clear_bit(node->qid, ht 150 u16 qid; local 292 int qid; local 349 u16 qid; local 431 mlx5e_htb_node_find_by_qid(struct mlx5e_htb *htb, u16 qid) argument 449 u16 qid, moved_qid; local 547 u16 qid; local [all...] |
H A D | qos.c | 38 u16 mlx5e_qid_from_qos(struct mlx5e_channels *chs, u16 qid) argument 49 return (chs->params.num_channels + is_ptp) * mlx5e_get_dcb_num_tc(&chs->params) + qid; 54 static struct mlx5e_txqsq *mlx5e_get_qos_sq(struct mlx5e_priv *priv, int qid) argument 61 ix = qid % params->num_channels; 62 qid /= params->num_channels; 66 return mlx5e_state_dereference(priv, qos_sqs[qid]); 76 int txq_ix, ix, qid, err = 0; local 113 qid = node_qid / params->num_channels; 139 rcu_assign_pointer(qos_sqs[qid], sq); 161 u16 qid; local 186 mlx5e_deactivate_qos_sq(struct mlx5e_priv *priv, u16 qid) argument 206 mlx5e_close_qos_sq(struct mlx5e_priv *priv, u16 qid) argument 335 u16 qid = params->num_channels * i + c->ix; local 358 mlx5e_reactivate_qos_sq(struct mlx5e_priv *priv, u16 qid, struct netdev_queue *txq) argument 365 mlx5e_reset_qdisc(struct net_device *dev, u16 qid) argument [all...] |
/linux-master/drivers/s390/crypto/ |
H A D | ap_queue.c | 56 status = ap_aqic(aq->qid, qirqctrl, virt_to_phys(ind)); 68 AP_QID_CARD(aq->qid), 69 AP_QID_QUEUE(aq->qid)); 80 * @qid: The AP queue number 92 __ap_send(ap_qid_t qid, unsigned long psmid, void *msg, size_t msglen, argument 96 qid |= 0x400000UL; 97 return ap_nqap(qid, psmid, msg, msglen); 131 status = ap_dqap(aq->qid, &aq->reply->psmid, 164 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); 244 ap_qid_t qid = aq->qid; local 1107 ap_queue_create(ap_qid_t qid, int device_type) argument [all...] |
/linux-master/drivers/vdpa/pds/ |
H A D | cmds.h | 13 int pds_vdpa_cmd_init_vq(struct pds_vdpa_device *pdsv, u16 qid, u16 invert_idx, 15 int pds_vdpa_cmd_reset_vq(struct pds_vdpa_device *pdsv, u16 qid, u16 invert_idx,
|
H A D | vdpa_dev.c | 72 static int pds_vdpa_set_vq_address(struct vdpa_device *vdpa_dev, u16 qid, argument 77 pdsv->vqs[qid].desc_addr = desc_addr; 78 pdsv->vqs[qid].avail_addr = driver_addr; 79 pdsv->vqs[qid].used_addr = device_addr; 84 static void pds_vdpa_set_vq_num(struct vdpa_device *vdpa_dev, u16 qid, u32 num) argument 88 pdsv->vqs[qid].q_len = num; 91 static void pds_vdpa_kick_vq(struct vdpa_device *vdpa_dev, u16 qid) argument 95 iowrite16(qid, pdsv->vqs[qid].notify); 98 static void pds_vdpa_set_vq_cb(struct vdpa_device *vdpa_dev, u16 qid, argument 117 pds_vdpa_release_irq(struct pds_vdpa_device *pdsv, int qid) argument 126 pds_vdpa_set_vq_ready(struct vdpa_device *vdpa_dev, u16 qid, bool ready) argument 164 pds_vdpa_get_vq_ready(struct vdpa_device *vdpa_dev, u16 qid) argument 171 pds_vdpa_set_vq_state(struct vdpa_device *vdpa_dev, u16 qid, const struct vdpa_vq_state *state) argument 221 pds_vdpa_get_vq_state(struct vdpa_device *vdpa_dev, u16 qid, struct vdpa_vq_state *state) argument 257 pds_vdpa_get_vq_notification(struct vdpa_device *vdpa_dev, u16 qid) argument 274 pds_vdpa_get_vq_irq(struct vdpa_device *vdpa_dev, u16 qid) argument 385 int max_vq, nintrs, qid, err; local 433 int qid; local 495 pds_vdpa_init_vqs_entry(struct pds_vdpa_device *pdsv, int qid, void __iomem *notify) argument [all...] |
H A D | cmds.c | 125 int pds_vdpa_cmd_init_vq(struct pds_vdpa_device *pdsv, u16 qid, u16 invert_idx, argument 134 .vdpa_vq_init.qid = cpu_to_le16(qid), 139 .vdpa_vq_init.intr_index = cpu_to_le16(qid), 146 dev_dbg(dev, "%s: qid %d len %d desc_addr %#llx avail_addr %#llx used_addr %#llx\n", 147 __func__, qid, ilog2(vq_info->q_len), 154 qid, comp.status, ERR_PTR(err)); 159 int pds_vdpa_cmd_reset_vq(struct pds_vdpa_device *pdsv, u16 qid, u16 invert_idx, argument 168 .vdpa_vq_reset.qid = cpu_to_le16(qid), [all...] |
/linux-master/drivers/vdpa/ifcvf/ |
H A D | ifcvf_base.c | 13 u16 ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector) argument 17 vp_iowrite16(qid, &cfg->queue_select); 72 u16 ifcvf_get_vq_size(struct ifcvf_hw *hw, u16 qid) argument 76 if (qid >= hw->nr_vring) 79 vp_iowrite16(qid, &hw->common_cfg->queue_select); 87 u16 queue_size, max_size, qid; local 90 for (qid = 1; qid < hw->nr_vring; qid++) { 91 queue_size = ifcvf_get_vq_size(hw, qid); 328 ifcvf_get_vq_state(struct ifcvf_hw *hw, u16 qid) argument 338 ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num) argument 347 ifcvf_set_vq_num(struct ifcvf_hw *hw, u16 qid, u32 num) argument 355 ifcvf_set_vq_address(struct ifcvf_hw *hw, u16 qid, u64 desc_area, u64 driver_area, u64 device_area) argument 371 ifcvf_get_vq_ready(struct ifcvf_hw *hw, u16 qid) argument 382 ifcvf_set_vq_ready(struct ifcvf_hw *hw, u16 qid, bool ready) argument 392 u16 qid; local 428 ifcvf_notify_queue(struct ifcvf_hw *hw, u16 qid) argument [all...] |
H A D | ifcvf_base.h | 108 void ifcvf_notify_queue(struct ifcvf_hw *hw, u16 qid); 120 u16 ifcvf_get_vq_state(struct ifcvf_hw *hw, u16 qid); 121 int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num); 125 u16 ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector); 127 void ifcvf_set_vq_num(struct ifcvf_hw *hw, u16 qid, u32 num); 128 int ifcvf_set_vq_address(struct ifcvf_hw *hw, u16 qid, u64 desc_area, 130 bool ifcvf_get_vq_ready(struct ifcvf_hw *hw, u16 qid); 131 void ifcvf_set_vq_ready(struct ifcvf_hw *hw, u16 qid, bool ready); 135 u16 ifcvf_get_vq_size(struct ifcvf_hw *hw, u16 qid);
|
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | resource.c | 111 u32 qid; local 119 qid = entry->qid; 122 qid = c4iw_get_resource(&rdev->resource.qid_table); 123 if (!qid) 126 rdev->stats.qid.cur += rdev->qpmask + 1; 128 for (i = qid+1; i & rdev->qpmask; i++) { 132 entry->qid = i; 143 entry->qid = qid; 163 c4iw_put_cqid(struct c4iw_rdev *rdev, u32 qid, struct c4iw_dev_ucontext *uctx) argument 181 u32 qid; local 237 c4iw_put_qpid(struct c4iw_rdev *rdev, u32 qid, struct c4iw_dev_ucontext *uctx) argument [all...] |
/linux-master/drivers/net/ethernet/marvell/prestera/ |
H A D | prestera_pci.c | 242 static u32 prestera_fw_evtq_len(struct prestera_fw *fw, u8 qid) argument 244 return fw->evt_queue[qid].len; 247 static u32 prestera_fw_evtq_avail(struct prestera_fw *fw, u8 qid) argument 249 u32 wr_idx = prestera_fw_read(fw, PRESTERA_EVTQ_WR_IDX_REG(qid)); 250 u32 rd_idx = prestera_fw_read(fw, PRESTERA_EVTQ_RD_IDX_REG(qid)); 252 return CIRC_CNT(wr_idx, rd_idx, prestera_fw_evtq_len(fw, qid)); 256 u8 qid, u32 idx) 258 u32 rd_idx = idx & (prestera_fw_evtq_len(fw, qid) - 1); 260 prestera_fw_write(fw, PRESTERA_EVTQ_RD_IDX_REG(qid), rd_idx); 263 static u8 __iomem *prestera_fw_evtq_buf(struct prestera_fw *fw, u8 qid) argument 255 prestera_fw_evtq_rd_set(struct prestera_fw *fw, u8 qid, u32 idx) argument 268 prestera_fw_evtq_read32(struct prestera_fw *fw, u8 qid) argument 278 prestera_fw_evtq_read_buf(struct prestera_fw *fw, u8 qid, void *buf, size_t len) argument 298 int qid; local 321 u8 qid; local 361 prestera_fw_cmdq_lock(struct prestera_fw *fw, u8 qid) argument 366 prestera_fw_cmdq_unlock(struct prestera_fw *fw, u8 qid) argument 371 prestera_fw_cmdq_len(struct prestera_fw *fw, u8 qid) argument 376 prestera_fw_cmdq_buf(struct prestera_fw *fw, u8 qid) argument 381 prestera_fw_cmd_send(struct prestera_fw *fw, int qid, void *in_msg, size_t in_size, void *out_msg, size_t out_size, unsigned int waitms) argument 434 prestera_fw_send_req(struct prestera_device *dev, int qid, void *in_msg, size_t in_size, void *out_msg, size_t out_size, unsigned int waitms) argument 455 u8 qid; local [all...] |
/linux-master/fs/quota/ |
H A D | kqid.c | 8 * @left: A qid value 11 * Return true if the two qid values are equal and false otherwise. 31 * qid_lt - Test to see if one qid value is less than another 32 * @left: The possibly lesser qid value 33 * @right: The possibly greater qid value 57 * from_kqid - Create a qid from a kqid user-namespace pair. 58 * @targ: The user namespace we want a qid in. 62 * return the resulting qid. 84 * from_kqid_munged - Create a qid from a kqid user-namespace pair. 85 * @targ: The user namespace we want a qid i 120 qid_valid(struct kqid qid) argument [all...] |
/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 398 "%d: [qid:%d, type:%d, subtype:%d, " 564 * @qid: Work queue identifier. 570 lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid) argument 575 if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid) 578 pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid); 583 if (phba->sli4_hba.els_wq->queue_id == qid) { 584 pr_err("ELS WQ[Qid:%d]\n", qid); 589 if (phba->sli4_hba.nvmels_wq->queue_id == qid) { 590 pr_err("NVME LS WQ[Qid:%d]\n", qid); 598 * @qid 604 lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid) argument 621 lpfc_debug_dump_rq_by_id(struct lpfc_hba *phba, int qid) argument 643 lpfc_debug_dump_cq_by_id(struct lpfc_hba *phba, int qid) argument 684 lpfc_debug_dump_eq_by_id(struct lpfc_hba *phba, int qid) argument [all...] |
/linux-master/drivers/nvme/target/ |
H A D | fabrics-cmd.c | 140 u16 qid = le16_to_cpu(c->qid); local 154 if (ctrl->sqs[qid] != NULL) { 155 pr_warn("qid %u has already been created\n", qid); 156 req->error_loc = offsetof(struct nvmf_connect_command, qid); 161 if (qid && sqsize > mqes) { 177 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); 178 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); 189 qid, ctr 291 u16 qid = le16_to_cpu(c->qid); local [all...] |
H A D | trace.h | 28 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ 31 (qid ? \ 73 __field(int, qid) 86 __entry->qid = req->sq->qid; 100 __entry->qid, __entry->cid, __entry->nsid, 102 show_opcode_name(__entry->qid, __entry->opcode, 104 parse_nvme_cmd(__entry->qid, __entry->opcode, 114 __field(int, qid) 121 __entry->qid [all...] |
H A D | fabrics-cmd-auth.c | 20 pr_debug("%s: ctrl %d qid %d transaction %u expired, resetting\n", 21 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid); 40 pr_debug("%s: ctrl %d qid %d: data sc_d %d napd %d authid %d halen %d dhlen %d\n", 41 __func__, ctrl->cntlid, req->sq->qid, 69 pr_debug("%s: ctrl %d qid %d: no usable hash found\n", 70 __func__, ctrl->cntlid, req->sq->qid); 73 pr_debug("%s: ctrl %d qid %d: no usable hash found, falling back to %s\n", 74 __func__, ctrl->cntlid, req->sq->qid, 97 pr_debug("%s: ctrl %d qid %d: no usable DH group found\n", 98 __func__, ctrl->cntlid, req->sq->qid); [all...] |
/linux-master/drivers/nvme/host/ |
H A D | trace.h | 26 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ 29 ((qid) ? \ 53 __field(int, qid) 64 __entry->qid = nvme_req_qid(req); 77 __entry->qid, __entry->cid, __entry->nsid, 79 show_opcode_name(__entry->qid, __entry->opcode, 81 parse_nvme_cmd(__entry->qid, __entry->opcode, 91 __field(int, qid) 100 __entry->qid = nvme_req_qid(req); 110 __entry->qid, __entr [all...] |
H A D | auth.c | 28 int qid; member in struct:nvme_dhchap_queue_context 57 static int nvme_auth_submit(struct nvme_ctrl *ctrl, int qid, argument 65 if (qid != 0) { 83 qid == 0 ? NVME_QID_ANY : qid, flags); 86 "qid %d auth_send failed with status %d\n", qid, ret); 89 "qid %d auth_send failed with error %d\n", qid, ret); 93 static int nvme_auth_receive_validate(struct nvme_ctrl *ctrl, int qid, argument 854 nvme_auth_negotiate(struct nvme_ctrl *ctrl, int qid) argument 875 nvme_auth_wait(struct nvme_ctrl *ctrl, int qid) argument [all...] |
/linux-master/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_xsk.h | 13 unsigned int qid) 15 return dp->xdp_prog && dp->xsk_pools[qid]; 12 nfp_net_has_xsk_pool_slow(struct nfp_net_dp *dp, unsigned int qid) argument
|
/linux-master/fs/xfs/ |
H A D | xfs_quotaops.c | 216 struct kqid qid, 225 id = from_kqid(&init_user_ns, qid); 226 return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); 229 /* Return quota info for active quota >= this qid */ 233 struct kqid *qid, 243 id = from_kqid(&init_user_ns, *qid); 244 ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), 250 *qid = make_kqid(current_user_ns(), qid->type, id); 257 struct kqid qid, 214 xfs_fs_get_dqblk( struct super_block *sb, struct kqid qid, struct qc_dqblk *qdq) argument 231 xfs_fs_get_nextdqblk( struct super_block *sb, struct kqid *qid, struct qc_dqblk *qdq) argument 255 xfs_fs_set_dqblk( struct super_block *sb, struct kqid qid, struct qc_dqblk *qdq) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_process_queue_manager.c | 33 struct process_queue_manager *pqm, unsigned int qid) 38 if ((pqn->q && pqn->q->properties.queue_id == qid) || 39 (pqn->kq && pqn->kq->queue->properties.queue_id == qid)) 47 unsigned int qid) 49 if (qid >= KFD_MAX_NUM_OF_QUEUES_PER_PROCESS) 52 if (__test_and_set_bit(qid, pqm->queue_slot_bitmap)) { 53 pr_err("Cannot create new queue because requested qid(%u) is in use\n", qid); 61 unsigned int *qid) 77 *qid 32 get_queue_by_qid( struct process_queue_manager *pqm, unsigned int qid) argument 46 assign_queue_slot_by_qid(struct process_queue_manager *pqm, unsigned int qid) argument 60 find_available_queue_slot(struct process_queue_manager *pqm, unsigned int *qid) argument 95 pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, void *gws) argument 223 init_user_queue(struct process_queue_manager *pqm, struct kfd_node *dev, struct queue **q, struct queue_properties *q_properties, struct file *f, struct amdgpu_bo *wptr_bo, unsigned int qid) argument 270 pqm_create_queue(struct process_queue_manager *pqm, struct kfd_node *dev, struct file *f, struct queue_properties *properties, unsigned int *qid, struct amdgpu_bo *wptr_bo, const struct kfd_criu_queue_priv_data *q_data, const void *restore_mqd, const void *restore_ctl_stack, uint32_t *p_doorbell_offset_in_process) argument 444 pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid) argument 511 pqm_update_queue_properties(struct process_queue_manager *pqm, unsigned int qid, struct queue_properties *p) argument 537 pqm_update_mqd(struct process_queue_manager *pqm, unsigned int qid, struct mqd_update_info *minfo) argument 579 pqm_get_kernel_queue( struct process_queue_manager *pqm, unsigned int qid) argument 592 pqm_get_user_queue(struct process_queue_manager *pqm, unsigned int qid) argument 601 pqm_get_wave_state(struct process_queue_manager *pqm, unsigned int qid, void __user *ctl_stack, u32 *ctl_stack_used_size, u32 *save_area_used_size) argument 720 pqm_checkpoint_mqd(struct process_queue_manager *pqm, unsigned int qid, void *mqd, void *ctl_stack) argument 986 pqm_get_queue_checkpoint_info(struct process_queue_manager *pqm, unsigned int qid, uint32_t *mqd_size, uint32_t *ctl_stack_size) argument [all...] |
/linux-master/include/linux/ |
H A D | quota.h | 79 extern qid_t from_kqid(struct user_namespace *to, struct kqid qid); 80 extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid); 81 extern bool qid_valid(struct kqid qid); 84 * make_kqid - Map a user-namespace, type, qid tuple into a kqid. 85 * @from: User namespace that the qid is in 87 * @qid: Quota identifier 89 * Maps a user-namespace, type qid tuple into a kernel internal 93 * qid tuple an invalid kqid is returned. Callers are expected to 98 enum quota_type type, qid_t qid) 105 kqid.uid = make_kuid(from, qid); 97 make_kqid(struct user_namespace *from, enum quota_type type, qid_t qid) argument 187 qid_has_mapping(struct user_namespace *ns, struct kqid qid) argument 514 quota_send_warning(struct kqid qid, dev_t dev, const char warntype) argument [all...] |