/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_hw_queues.c | 314 struct hw_rq *rq = NULL; local 324 * encapsulates 2 SLI queues (for rq pair) 327 rq = kzalloc(sizeof(*rq), GFP_KERNEL); 328 if (!rq) 331 rqs[i] = rq; 332 rq->instance = hw->hw_rq_count++; 333 rq->cq = cqs[i]; 334 rq->type = SLI4_QTYPE_RQ; 335 rq 455 efct_hw_del_rq(struct hw_rq *rq) argument 494 struct hw_rq *rq = hw->hw_rq[hw->hw_rq_lookup[rqindex]]; local 531 struct hw_rq *rq; local 607 struct hw_rq *rq = hw->hw_rq[hw_rq_index]; local [all...] |
/linux-master/block/ |
H A D | blk-rq-qos.c | 3 #include "blk-rq-qos.h" 35 void __rq_qos_done(struct rq_qos *rqos, struct request *rq) argument 39 rqos->ops->done(rqos, rq); 44 void __rq_qos_issue(struct rq_qos *rqos, struct request *rq) argument 48 rqos->ops->issue(rqos, rq); 53 void __rq_qos_requeue(struct rq_qos *rqos, struct request *rq) argument 57 rqos->ops->requeue(rqos, rq); 71 void __rq_qos_track(struct rq_qos *rqos, struct request *rq, struct bio *bio) argument 75 rqos->ops->track(rqos, rq, bio); 80 void __rq_qos_merge(struct rq_qos *rqos, struct request *rq, struc argument [all...] |
H A D | blk-wbt.c | 30 #include "blk-rq-qos.h" 101 static inline void wbt_clear_state(struct request *rq) argument 103 rq->wbt_flags = 0; 106 static inline enum wbt_flags wbt_flags(struct request *rq) argument 108 return rq->wbt_flags; 111 static inline bool wbt_is_tracked(struct request *rq) argument 113 return rq->wbt_flags & WBT_TRACKED; 116 static inline bool wbt_is_read(struct request *rq) argument 118 return rq->wbt_flags & WBT_READ; 245 static void wbt_done(struct rq_qos *rqos, struct request *rq) argument 664 wbt_track(struct rq_qos *rqos, struct request *rq, struct bio *bio) argument 670 wbt_issue(struct rq_qos *rqos, struct request *rq) argument 690 wbt_requeue(struct rq_qos *rqos, struct request *rq) argument 743 wbt_data_dir(const struct request *rq) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_rc6.c | 127 struct i915_request *rq; local 132 rq = intel_context_create_request(ce); 133 if (IS_ERR(rq)) 134 return ERR_CAST(rq); 136 cs = intel_ring_begin(rq, 4); 138 i915_request_add(rq); 143 if (GRAPHICS_VER(rq->i915) >= 8) 150 intel_ring_advance(rq, cs); 152 result = rq->hwsp_seqno + 2; 153 i915_request_add(rq); [all...] |
H A D | selftest_rps.c | 246 struct i915_request *rq; local 255 rq = igt_spinner_create_request(&spin, 258 if (IS_ERR(rq)) { 260 err = PTR_ERR(rq); 264 i915_request_add(rq); 266 if (!igt_wait_for_spinner(&spin, rq)) { 405 struct i915_request *rq; local 415 rq = igt_spinner_create_request(&spin, 418 if (IS_ERR(rq)) { 419 err = PTR_ERR(rq); 633 struct i915_request *rq; local 772 struct i915_request *rq; local 899 struct i915_request *rq; local 1161 struct i915_request *rq; local 1263 struct i915_request *rq; local [all...] |
H A D | intel_engine_pm.h | 91 struct i915_request *rq; local 102 rq = i915_request_create(engine->kernel_context); 105 return rq;
|
H A D | intel_workarounds.h | 25 int intel_engine_emit_ctx_wa(struct i915_request *rq);
|
H A D | intel_execlists_submission.h | 30 const struct i915_request *rq,
|
H A D | intel_gt_requests.c | 19 struct i915_request *rq, *rn; local 21 list_for_each_entry_safe(rq, rn, &tl->requests, link) 22 if (!i915_request_retire(rq)) 243 struct i915_request *rq, *rn; local 250 llist_for_each_entry_safe(rq, rn, first, watchdog.link) { 251 if (!i915_request_completed(rq)) { 252 struct dma_fence *f = &rq->fence; 258 i915_request_cancel(rq, -EINTR); 260 i915_request_put(rq);
|
H A D | mock_engine.c | 256 static void mock_add_to_engine(struct i915_request *rq) argument 258 lockdep_assert_held(&rq->engine->sched_engine->lock); 259 list_move_tail(&rq->sched.link, &rq->engine->sched_engine->requests); 262 static void mock_remove_from_engine(struct i915_request *rq) argument 268 * as their rq->engine pointer is not stable until under that 270 * check that the rq still belongs to the newly locked engine. 273 locked = READ_ONCE(rq->engine); 275 while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) { 280 list_del_init(&rq 297 struct i915_request *rq; local [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_debugfs.c | 53 static u64 hinic_dbg_get_rq_info(struct hinic_dev *nic_dev, struct hinic_rq *rq, int idx) argument 55 struct hinic_wq *wq = rq->wq; 59 return nic_dev->hwdev->func_to_io.global_qpn + rq->qid; 61 return be16_to_cpu(*(__be16 *)(rq->pi_virt_addr)) & wq->mask; 67 return rq->msix_entry; 236 struct hinic_rq *rq; local 240 rq = dev->rxqs[rq_id].rq; 246 return create_dbg_files(dev, HINIC_DBG_RQ_INFO, rq, root, &rq 250 hinic_rq_debug_rem(struct hinic_rq *rq) argument [all...] |
/linux-master/drivers/isdn/mISDN/ |
H A D | stack.c | 421 struct channel_req rq; local 437 rq.protocol = protocol; 438 rq.adr.channel = adr->channel; 439 err = dev->D.ctrl(&dev->D, OPEN_CHANNEL, &rq); 458 struct channel_req rq, rq2; local 470 rq.protocol = protocol; 471 rq.adr = *adr; 472 err = dev->D.ctrl(&dev->D, OPEN_CHANNEL, &rq); 475 ch->recv = rq.ch->send; 476 ch->peer = rq 515 struct channel_req rq; local [all...] |
/linux-master/tools/testing/selftests/rcutorture/bin/ |
H A D | kvm-transform.sh | 103 rq = ""; 111 rq = substr(arg, length($i), 1); 131 line = line rq;
|
/linux-master/drivers/block/rnbd/ |
H A D | rnbd-proto.h | 257 static inline u32 rq_to_rnbd_flags(struct request *rq) argument 261 switch (req_op(rq)) { 282 (__force u32)req_op(rq), 283 (__force unsigned long long)rq->cmd_flags); 287 if (op_is_sync(rq->cmd_flags)) 290 if (op_is_flush(rq->cmd_flags))
|
/linux-master/drivers/scsi/ |
H A D | scsi_ioctl.c | 348 static int scsi_fill_sghdr_rq(struct scsi_device *sdev, struct request *rq, argument 351 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); 361 rq->timeout = msecs_to_jiffies(hdr->timeout); 362 if (!rq->timeout) 363 rq->timeout = sdev->sg_timeout; 364 if (!rq->timeout) 365 rq->timeout = BLK_DEFAULT_SG_TIMEOUT; 366 if (rq->timeout < BLK_MIN_SG_TIMEOUT) 367 rq->timeout = BLK_MIN_SG_TIMEOUT; 372 static int scsi_complete_sghdr_rq(struct request *rq, struc argument 417 struct request *rq; local 507 struct request *rq; local [all...] |
H A D | scsi_debugfs.c | 54 void scsi_show_rq(struct seq_file *m, struct request *rq) argument 56 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq); 58 int timeout_ms = jiffies_to_msecs(rq->timeout);
|
/linux-master/drivers/platform/chrome/wilco_ec/ |
H A D | telemetry.c | 142 * @rq: Request buffer copied from userspace. 154 static int check_telem_request(struct wilco_ec_telem_request *rq, argument 159 if (rq->reserved) 162 switch (rq->command) { 164 max_size += sizeof(rq->args.get_log); 167 max_size += sizeof(rq->args.get_version); 170 max_size += sizeof(rq->args.get_fan_info); 173 max_size += sizeof(rq->args.get_diag_info); 176 max_size += sizeof(rq->args.get_temp_info); 179 max_size += sizeof(rq [all...] |
/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_ptp.c | 164 struct ptp_clock_request *rq, int on) 174 switch (rq->type) { 177 if (rq->perout.flags) 180 cfg = &priv->pps[rq->perout.index]; 182 cfg->start.tv_sec = rq->perout.start.sec; 183 cfg->start.tv_nsec = rq->perout.start.nsec; 184 cfg->period.tv_sec = rq->perout.period.sec; 185 cfg->period.tv_nsec = rq->perout.period.nsec; 189 rq->perout.index, cfg, on, 206 rq 163 stmmac_enable(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 384 request_to_scheduling_context(struct i915_request *rq) argument 386 return intel_context_to_parent(rq->context); 696 static int __guc_add_request(struct intel_guc *guc, struct i915_request *rq) argument 699 struct intel_context *ce = request_to_scheduling_context(rq); 705 lockdep_assert_held(&rq->engine->sched_engine->lock); 712 i915_request_put(i915_request_mark_eio(rq)); 772 trace_i915_request_guc_submit(rq); 779 static int guc_add_request(struct intel_guc *guc, struct i915_request *rq) argument 781 int ret = __guc_add_request(guc, rq); 784 guc->stalled_request = rq; 791 guc_set_lrc_tail(struct i915_request *rq) argument 797 rq_prio(const struct i915_request *rq) argument 802 is_multi_lrc_rq(struct i915_request *rq) argument 807 can_merge_rq(struct i915_request *rq, struct i915_request *last) argument 850 __guc_wq_item_append(struct i915_request *rq) argument 893 guc_wq_item_append(struct intel_guc *guc, struct i915_request *rq) argument 911 multi_lrc_submit(struct i915_request *rq) argument 955 struct i915_request *rq, *rn; local 1770 struct i915_request *rq, *rn; local 1806 struct i915_request *rq; local 1908 struct i915_request *rq; local 1923 struct i915_request *rq, *rn; local 2142 queue_request(struct i915_sched_engine *sched_engine, struct i915_request *rq, int prio) argument 2153 guc_bypass_tasklet_submit(struct intel_guc *guc, struct i915_request *rq) argument 2179 need_tasklet(struct intel_guc *guc, struct i915_request *rq) argument 2189 guc_submit_request(struct i915_request *rq) argument 3151 guc_context_cancel_request(struct intel_context *ce, struct i915_request *rq) argument 3196 guc_context_revoke(struct intel_context *ce, struct i915_request *rq, unsigned int preempt_timeout_ms) argument 3652 add_to_context(struct i915_request *rq) argument 3676 guc_prio_fini(struct i915_request *rq, struct intel_context *ce) argument 3688 remove_from_context(struct i915_request *rq) argument 3741 struct i915_request *rq = container_of(wrk, typeof(*rq), submit_work); local 3749 struct i915_request *rq, *rn; local 3807 guc_request_alloc(struct i915_request *rq) argument 4263 guc_bump_inflight_request_prio(struct i915_request *rq, int prio) argument 4287 guc_retire_inflight_request_prio(struct i915_request *rq) argument 5319 struct i915_request *rq; local 5440 struct i915_request *rq; local 5555 emit_bb_start_parent_no_preempt_mid_batch(struct i915_request *rq, u64 offset, u32 len, const unsigned int flags) argument 5602 emit_bb_start_child_no_preempt_mid_batch(struct i915_request *rq, u64 offset, u32 len, const unsigned int flags) argument 5647 __emit_fini_breadcrumb_parent_no_preempt_mid_batch(struct i915_request *rq, u32 *cs) argument 5690 skip_handshake(struct i915_request *rq) argument 5697 emit_fini_breadcrumb_parent_no_preempt_mid_batch(struct i915_request *rq, u32 *cs) argument 5741 __emit_fini_breadcrumb_child_no_preempt_mid_batch(struct i915_request *rq, u32 *cs) argument 5773 emit_fini_breadcrumb_child_no_preempt_mid_batch(struct i915_request *rq, u32 *cs) argument [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | trap.h | 12 struct mlx5e_rq rq; member in struct:mlx5e_trap
|
H A D | dim.h | 42 int mlx5e_dim_rx_change(struct mlx5e_rq *rq, bool enabled);
|
H A D | ptp.c | 267 struct mlx5e_rq *rq = &c->rq; local 283 work_done = mlx5e_poll_rx_cq(&rq->cq, budget); 285 busy |= INDIRECT_CALL_2(rq->post_wqes, 288 rq); 308 mlx5e_cq_arm(&rq->cq); 608 struct mlx5e_cq *cq = &c->rq.cq; 689 struct mlx5e_rq *rq) 695 rq->wq_type = params->rq_wq_type; 696 rq 688 mlx5e_init_ptp_rq(struct mlx5e_ptp *c, struct mlx5e_params *params, struct mlx5e_rq *rq) argument [all...] |
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | igt_spinner.h | 41 bool igt_wait_for_spinner(struct igt_spinner *spin, struct i915_request *rq);
|
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_fp.c | 83 if (!qp->rq.flushed) { 87 qp->rq.flushed = true; 130 if (qp->rq.flushed) { 131 qp->rq.flushed = false; 146 qp->rq.hwq.prod = 0; 147 qp->rq.hwq.cons = 0; 177 struct bnxt_qplib_q *rq = &qp->rq; local 182 rq->max_wqe * qp->rq_hdr_buf_size, 199 struct bnxt_qplib_q *rq local 833 struct bnxt_qplib_q *rq = &qp->rq; local 976 struct bnxt_qplib_q *rq = &qp->rq; local 1578 struct bnxt_qplib_q *rq = &qp->rq; local 1591 struct bnxt_qplib_q *rq = &qp->rq; local 2039 struct bnxt_qplib_q *rq = &qp->rq; local 2048 struct bnxt_qplib_q *rq = &qp->rq; local 2314 __flush_rq(struct bnxt_qplib_q *rq, struct bnxt_qplib_qp *qp, struct bnxt_qplib_cqe **pcqe, int *budget) argument 2591 struct bnxt_qplib_q *rq; local 2671 struct bnxt_qplib_q *rq; local 2767 struct bnxt_qplib_q *rq; local 2860 struct bnxt_qplib_q *sq, *rq; local [all...] |
/linux-master/drivers/scsi/esas2r/ |
H A D | esas2r_init.c | 104 struct esas2r_request *rq) 126 rq->vrq_md = memdesc; 127 rq->vrq = (union atto_vda_req *)memdesc->virt_addr; 128 rq->vrq->scsi.handle = a->num_vrqs; 770 struct esas2r_request *rq; local 917 for (rq = a->first_ae_req, i = 0; i < num_ae_requests; rq++, 919 INIT_LIST_HEAD(&rq->req_list); 920 if (!alloc_vda_req(a, rq)) { 926 esas2r_rq_init_request(rq, 103 alloc_vda_req(struct esas2r_adapter *a, struct esas2r_request *rq) argument 1145 esas2r_format_init_msg(struct esas2r_adapter *a, struct esas2r_request *rq) argument 1246 struct esas2r_request *rq = &a->general_req; local 1287 struct esas2r_request *rq; local [all...] |