/linux-master/block/ |
H A D | blk-pm.h | 21 if (rq->q->dev && !(rq->rq_flags & RQF_PM))
|
H A D | blk-mq-sched.h | 40 if (rq->rq_flags & RQF_USE_SCHED) { 51 if (rq->rq_flags & RQF_USE_SCHED) { 61 if (rq->rq_flags & RQF_USE_SCHED) {
|
H A D | blk-zoned.c | 73 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); 74 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; 86 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); 87 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; 93 rq->rq_flags &= ~RQF_ZONE_WRITE_LOCKED;
|
H A D | blk-flush.c | 135 rq->rq_flags &= ~RQF_FLUSH_SEQ; 333 flush_rq->rq_flags |= RQF_FLUSH_SEQ; 386 rq->rq_flags |= RQF_FLUSH_SEQ;
|
H A D | blk-timeout.c | 140 req->rq_flags &= ~RQF_TIMED_OUT;
|
H A D | blk-mq.c | 360 data->rq_flags |= RQF_PM; 362 data->rq_flags |= RQF_IO_STAT; 363 rq->rq_flags = data->rq_flags; 365 if (data->rq_flags & RQF_SCHED_TAGS) { 390 if (rq->rq_flags & RQF_USE_SCHED) { 427 if (!(data->rq_flags & RQF_SCHED_TAGS)) 455 data->rq_flags |= RQF_SCHED_TAGS; 467 data->rq_flags |= RQF_USE_SCHED; 476 if (!(data->rq_flags [all...] |
H A D | blk-mq.h | 152 req_flags_t rq_flags; member in struct:blk_mq_alloc_data 224 if (data->rq_flags & RQF_SCHED_TAGS)
|
H A D | blk-merge.c | 568 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) 734 if (rq->rq_flags & RQF_MIXED_MERGE) 747 rq->rq_flags |= RQF_MIXED_MERGE; 766 if (req->rq_flags & RQF_MIXED_MERGE) { 847 if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
|
H A D | blk.h | 154 if (rq->rq_flags & RQF_NOMERGE_FLAGS) 286 #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED) 356 return (rq->rq_flags & RQF_IO_STAT) && !blk_rq_is_passthrough(rq);
|
/linux-master/include/linux/ |
H A D | blk-mq.h | 85 req_flags_t rq_flags; member in struct:request 840 return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS | RQF_USE_SCHED)); 845 return rq->rq_flags & RQF_RESV; 860 if (!iob || (req->rq_flags & RQF_SCHED_TAGS) || ioerror || 1078 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) 1089 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) 1128 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) 1189 if (rq->rq_flags & RQF_ZONE_WRITE_LOCKED)
|
/linux-master/kernel/sched/ |
H A D | stop_task.c | 19 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
|
H A D | sched.h | 1592 struct rq_flags { struct 1613 * copy of the (on-stack) 'struct rq_flags rf'. 1617 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf) 1630 static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf) 1640 static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf) 1652 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) 1655 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) 1659 static inline void __task_rq_unlock(struct rq *rq, struct rq_flags *rf) 1667 task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf) 1679 struct rq *rq; struct rq_flags r [all...] |
H A D | core.c | 628 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) 652 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) 791 struct rq_flags rf; 819 struct rq_flags rf; 1753 struct rq_flags rf; 2294 struct rq_flags rf; 2520 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, 2566 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, 2590 struct rq_flags rf; 2967 static int affine_move_task(struct rq *rq, struct task_struct *p, struct rq_flags *r [all...] |
H A D | core_sched.c | 57 struct rq_flags rf;
|
H A D | stats.h | 169 struct rq_flags rf;
|
H A D | idle.c | 445 balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
|
/linux-master/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_queue.h | 69 u16 rq_flags; member in struct:fun_queue 120 u16 rq_flags; member in struct:fun_queue_alloc_req
|
H A D | fun_queue.c | 459 funq->rq_flags = req->rq_flags | FUN_ADMIN_EPSQ_CREATE_FLAG_RQ; 528 rc = fun_sq_create(fdev, funq->rq_flags, funq->rqid, funq->cqid, 0,
|
/linux-master/include/linux/sunrpc/ |
H A D | svc.h | 207 unsigned long rq_flags; /* flags field */ member in struct:svc_rqst 240 /* bits for rq_flags */ 302 set_bit(RQ_VICTIM, &rqstp->rq_flags); 304 return test_bit(RQ_VICTIM, &rqstp->rq_flags);
|
/linux-master/drivers/mmc/core/ |
H A D | queue.c | 243 req->rq_flags |= RQF_QUIET; 291 if (!(req->rq_flags & RQF_DONTPREP)) { 293 req->rq_flags |= RQF_DONTPREP;
|
/linux-master/drivers/nvme/host/ |
H A D | ioctl.c | 101 struct nvme_command *cmd, blk_opf_t rq_flags, 106 req = blk_mq_alloc_request(q, nvme_req_op(cmd) | rq_flags, blk_flags); 446 blk_opf_t rq_flags = REQ_ALLOC_CACHE; local 481 rq_flags |= REQ_NOWAIT; 485 rq_flags |= REQ_POLLED; 487 req = nvme_alloc_user_request(q, &c, rq_flags, blk_flags); 100 nvme_alloc_user_request(struct request_queue *q, struct nvme_command *cmd, blk_opf_t rq_flags, blk_mq_req_flags_t blk_flags) argument
|
/linux-master/drivers/scsi/ |
H A D | scsi_lib.c | 118 if (rq->rq_flags & RQF_DONTPREP) { 119 rq->rq_flags &= ~RQF_DONTPREP; 320 req->rq_flags |= RQF_QUIET; 732 if (!(rq->rq_flags & RQF_MIXED_MERGE)) 911 if (!(req->rq_flags & RQF_QUIET)) { 1002 else if (req->rq_flags & RQF_QUIET) 1242 if (rq->rq_flags & RQF_DONTPREP) { 1244 rq->rq_flags &= ~RQF_DONTPREP; 1323 if (req && WARN_ON_ONCE(!(req->rq_flags & RQF_PM))) 1331 if (req && !(req->rq_flags [all...] |
/linux-master/drivers/md/ |
H A D | dm-rq.c | 266 if (rq->rq_flags & RQF_FAILED) 293 rq->rq_flags |= RQF_FAILED;
|
/linux-master/drivers/scsi/device_handler/ |
H A D | scsi_dh_hp_sw.c | 191 req->rq_flags |= RQF_QUIET;
|
/linux-master/net/sunrpc/ |
H A D | svc_xprt.c | 402 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) { 406 set_bit(RQ_DATA, &rqstp->rq_flags); 414 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) { 1177 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags)) 1208 set_bit(RQ_DROPME, &rqstp->rq_flags);
|