Searched refs:rq_flags (Results 1 - 25 of 61) sorted by relevance

123

/linux-master/block/
H A Dblk-pm.h21 if (rq->q->dev && !(rq->rq_flags & RQF_PM))
H A Dblk-mq-sched.h40 if (rq->rq_flags & RQF_USE_SCHED) {
51 if (rq->rq_flags & RQF_USE_SCHED) {
61 if (rq->rq_flags & RQF_USE_SCHED) {
H A Dblk-zoned.c73 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED);
74 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED;
86 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED);
87 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED;
93 rq->rq_flags &= ~RQF_ZONE_WRITE_LOCKED;
H A Dblk-flush.c135 rq->rq_flags &= ~RQF_FLUSH_SEQ;
333 flush_rq->rq_flags |= RQF_FLUSH_SEQ;
386 rq->rq_flags |= RQF_FLUSH_SEQ;
H A Dblk-timeout.c140 req->rq_flags &= ~RQF_TIMED_OUT;
H A Dblk-mq.c360 data->rq_flags |= RQF_PM;
362 data->rq_flags |= RQF_IO_STAT;
363 rq->rq_flags = data->rq_flags;
365 if (data->rq_flags & RQF_SCHED_TAGS) {
390 if (rq->rq_flags & RQF_USE_SCHED) {
427 if (!(data->rq_flags & RQF_SCHED_TAGS))
455 data->rq_flags |= RQF_SCHED_TAGS;
467 data->rq_flags |= RQF_USE_SCHED;
476 if (!(data->rq_flags
[all...]
H A Dblk-mq.h152 req_flags_t rq_flags; member in struct:blk_mq_alloc_data
224 if (data->rq_flags & RQF_SCHED_TAGS)
H A Dblk-merge.c568 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
734 if (rq->rq_flags & RQF_MIXED_MERGE)
747 rq->rq_flags |= RQF_MIXED_MERGE;
766 if (req->rq_flags & RQF_MIXED_MERGE) {
847 if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
H A Dblk.h154 if (rq->rq_flags & RQF_NOMERGE_FLAGS)
286 #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
356 return (rq->rq_flags & RQF_IO_STAT) && !blk_rq_is_passthrough(rq);
/linux-master/include/linux/
H A Dblk-mq.h85 req_flags_t rq_flags; member in struct:request
840 return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS | RQF_USE_SCHED));
845 return rq->rq_flags & RQF_RESV;
860 if (!iob || (req->rq_flags & RQF_SCHED_TAGS) || ioerror ||
1078 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
1089 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
1128 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
1189 if (rq->rq_flags & RQF_ZONE_WRITE_LOCKED)
/linux-master/kernel/sched/
H A Dstop_task.c19 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
H A Dsched.h1592 struct rq_flags { struct
1613 * copy of the (on-stack) 'struct rq_flags rf'.
1617 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf)
1630 static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf)
1640 static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf)
1652 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1655 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1659 static inline void __task_rq_unlock(struct rq *rq, struct rq_flags *rf)
1667 task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
1679 struct rq *rq; struct rq_flags r
[all...]
H A Dcore.c628 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf)
652 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf)
791 struct rq_flags rf;
819 struct rq_flags rf;
1753 struct rq_flags rf;
2294 struct rq_flags rf;
2520 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf,
2566 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf,
2590 struct rq_flags rf;
2967 static int affine_move_task(struct rq *rq, struct task_struct *p, struct rq_flags *r
[all...]
H A Dcore_sched.c57 struct rq_flags rf;
H A Dstats.h169 struct rq_flags rf;
H A Didle.c445 balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
/linux-master/drivers/net/ethernet/fungible/funcore/
H A Dfun_queue.h69 u16 rq_flags; member in struct:fun_queue
120 u16 rq_flags; member in struct:fun_queue_alloc_req
H A Dfun_queue.c459 funq->rq_flags = req->rq_flags | FUN_ADMIN_EPSQ_CREATE_FLAG_RQ;
528 rc = fun_sq_create(fdev, funq->rq_flags, funq->rqid, funq->cqid, 0,
/linux-master/include/linux/sunrpc/
H A Dsvc.h207 unsigned long rq_flags; /* flags field */ member in struct:svc_rqst
240 /* bits for rq_flags */
302 set_bit(RQ_VICTIM, &rqstp->rq_flags);
304 return test_bit(RQ_VICTIM, &rqstp->rq_flags);
/linux-master/drivers/mmc/core/
H A Dqueue.c243 req->rq_flags |= RQF_QUIET;
291 if (!(req->rq_flags & RQF_DONTPREP)) {
293 req->rq_flags |= RQF_DONTPREP;
/linux-master/drivers/nvme/host/
H A Dioctl.c101 struct nvme_command *cmd, blk_opf_t rq_flags,
106 req = blk_mq_alloc_request(q, nvme_req_op(cmd) | rq_flags, blk_flags);
446 blk_opf_t rq_flags = REQ_ALLOC_CACHE; local
481 rq_flags |= REQ_NOWAIT;
485 rq_flags |= REQ_POLLED;
487 req = nvme_alloc_user_request(q, &c, rq_flags, blk_flags);
100 nvme_alloc_user_request(struct request_queue *q, struct nvme_command *cmd, blk_opf_t rq_flags, blk_mq_req_flags_t blk_flags) argument
/linux-master/drivers/scsi/
H A Dscsi_lib.c118 if (rq->rq_flags & RQF_DONTPREP) {
119 rq->rq_flags &= ~RQF_DONTPREP;
320 req->rq_flags |= RQF_QUIET;
732 if (!(rq->rq_flags & RQF_MIXED_MERGE))
911 if (!(req->rq_flags & RQF_QUIET)) {
1002 else if (req->rq_flags & RQF_QUIET)
1242 if (rq->rq_flags & RQF_DONTPREP) {
1244 rq->rq_flags &= ~RQF_DONTPREP;
1323 if (req && WARN_ON_ONCE(!(req->rq_flags & RQF_PM)))
1331 if (req && !(req->rq_flags
[all...]
/linux-master/drivers/md/
H A Ddm-rq.c266 if (rq->rq_flags & RQF_FAILED)
293 rq->rq_flags |= RQF_FAILED;
/linux-master/drivers/scsi/device_handler/
H A Dscsi_dh_hp_sw.c191 req->rq_flags |= RQF_QUIET;
/linux-master/net/sunrpc/
H A Dsvc_xprt.c402 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) {
406 set_bit(RQ_DATA, &rqstp->rq_flags);
414 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) {
1177 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags))
1208 set_bit(RQ_DROPME, &rqstp->rq_flags);

Completed in 267 milliseconds

123