/linux-master/include/linux/ |
H A D | t10-pi.h | 48 return blk_rq_pos(rq) >> (shift - SECTOR_SHIFT) & 0xffffffff; 79 return lower_48_bits(blk_rq_pos(rq) >> (shift - SECTOR_SHIFT));
|
H A D | blktrace_api.h | 121 if (blk_rq_is_passthrough(rq) || blk_rq_pos(rq) == (sector_t)-1) 123 return blk_rq_pos(rq);
|
H A D | blk-mq.h | 1029 * blk_rq_pos() : the current sector 1036 static inline sector_t blk_rq_pos(const struct request *rq) function 1156 return disk_zone_no(rq->q->disk, blk_rq_pos(rq)); 1161 return disk_zone_is_seq(rq->q->disk, blk_rq_pos(rq));
|
/linux-master/block/ |
H A D | elevator.c | 54 #define rq_hash_key(rq) (blk_rq_pos(rq) + blk_rq_sectors(rq)) 237 if (blk_rq_pos(rq) < blk_rq_pos(__rq)) 239 else if (blk_rq_pos(rq) >= blk_rq_pos(__rq)) 264 if (sector < blk_rq_pos(rq)) 266 else if (sector > blk_rq_pos(rq)) 357 __rq = elv_rqhash_find(q, blk_rq_pos(rq));
|
H A D | blk-merge.c | 642 blk_rq_get_max_sectors(req, blk_rq_pos(req))) { 677 blk_rq_get_max_sectors(req, blk_rq_pos(req))) 699 blk_rq_get_max_sectors(req, blk_rq_pos(req))) 791 else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next)) 958 else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector) 960 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector) 1043 blk_rq_get_max_sectors(req, blk_rq_pos(req)))
|
H A D | mq-deadline.c | 160 * Return the first request for which blk_rq_pos() >= @pos. For zoned devices, 183 if (blk_rq_pos(rq) >= pos) { 321 return blk_rq_pos(prev) + blk_rq_sectors(prev) == blk_rq_pos(rq); 331 sector_t pos = blk_rq_pos(rq); 336 } while (rq && blk_rq_pos(rq) == pos); 370 rb_rq = deadline_from_pos(per_prio, data_dir, blk_rq_pos(rq)); 371 if (rb_rq && blk_rq_pos(rb_rq) < blk_rq_pos(rq)) 550 dd->per_prio[prio].latest_pos[data_dir] = blk_rq_pos(r [all...] |
H A D | bfq-iosched.c | 518 s1 = blk_rq_pos(rq1); 519 s2 = blk_rq_pos(rq2); 747 if (sector > blk_rq_pos(bfqq->next_rq)) 749 else if (sector < blk_rq_pos(bfqq->next_rq)) 813 blk_rq_pos(bfqq->next_rq), &parent, &p); 1039 return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last)); 2375 return abs(blk_rq_pos(rq) - last_pos); 2508 blk_rq_pos(req) < 2509 blk_rq_pos(container_of(rb_prev(&req->rb_node), 2669 return blk_rq_pos(io_struc [all...] |
/linux-master/include/scsi/ |
H A D | scsi_cmnd.h | 225 return blk_rq_pos(scsi_cmd_to_rq(scmd)); 232 return blk_rq_pos(scsi_cmd_to_rq(scmd)) >> shift;
|
/linux-master/drivers/block/ |
H A D | z2ram.c | 72 unsigned long start = blk_rq_pos(req) << 9; 80 (unsigned long long)blk_rq_pos(req),
|
H A D | virtio_blk.c | 185 range[0].sector = cpu_to_le64(blk_rq_pos(req)); 257 sector = blk_rq_pos(req); 261 sector = blk_rq_pos(req); 278 sector = blk_rq_pos(req); 282 sector = blk_rq_pos(req); 286 sector = blk_rq_pos(req); 290 sector = blk_rq_pos(req); 295 sector = blk_rq_pos(req);
|
H A D | swim3.c | 343 * Do not remove the cast. blk_rq_pos(req) is now a sector_t and can be 347 fs->req_cyl = ((long)blk_rq_pos(req)) / fs->secpercyl; 348 x = ((long)blk_rq_pos(req)) % fs->secpercyl; 649 (long)blk_rq_pos(fs->cur_req)); 767 (long)blk_rq_pos(req), err);
|
H A D | ps3disk.c | 116 start_sector = blk_rq_pos(req) * priv->blocking_factor;
|
/linux-master/drivers/md/ |
H A D | dm-rq.c | 134 blk_rq_pos(orig), tio->n_sectors, true, 385 blk_rq_pos(rq)); 443 blk_rq_pos(orig), tio->n_sectors, false, 0,
|
/linux-master/drivers/s390/block/ |
H A D | dasd_fba.c | 333 first_rec = blk_rq_pos(req) >> block->s2b_shift; 335 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; 450 first_rec = blk_rq_pos(req) >> block->s2b_shift; 452 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; 488 block->bp_block, blk_rq_pos(req), blk_rq_sectors(req));
|
H A D | dasd_diag.c | 526 first_rec = blk_rq_pos(req) >> block->s2b_shift; 528 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
|
H A D | scm_blk.c | 196 msb->scm_addr = scmdev->address + ((u64) blk_rq_pos(req) << 9);
|
/linux-master/include/trace/events/ |
H A D | block.h | 122 __entry->sector = blk_rq_pos(rq); 567 __entry->sector = blk_rq_pos(rq);
|
/linux-master/drivers/scsi/ |
H A D | sd_zbc.c | 334 sector_t sector = blk_rq_pos(rq); 481 sector_t sector = blk_rq_pos(rq);
|
H A D | sr.c | 332 good_bytes = (error_sector - blk_rq_pos(rq)) << 9; 431 if (((unsigned int)blk_rq_pos(rq) % (s_size >> 9)) || 447 block = (unsigned int)blk_rq_pos(rq) / (s_size >> 9);
|
H A D | sd.c | 895 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq)); 926 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq)); 953 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq)); 979 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq)); 1216 sector_t lba = sectors_to_logical(sdp, blk_rq_pos(rq)); 1237 if (blk_rq_pos(rq) + blk_rq_sectors(rq) > get_capacity(rq->q->disk)) { 1242 if ((blk_rq_pos(rq) & mask) || (blk_rq_sectors(rq) & mask)) { 1310 (unsigned long long)blk_rq_pos(rq), 2144 start_lba = sectors_to_logical(sdev, blk_rq_pos(req));
|
/linux-master/drivers/nvme/host/ |
H A D | zns.c | 244 c->zms.slba = cpu_to_le64(nvme_sect_to_lba(ns->head, blk_rq_pos(req)));
|
/linux-master/drivers/mmc/core/ |
H A D | block.c | 1165 from = blk_rq_pos(req); 1220 from = blk_rq_pos(req); 1299 if (!IS_ALIGNED(blk_rq_pos(req), card->ext_csd.rel_sectors)) 1401 brq->data.blk_addr = blk_rq_pos(req); 1425 (blk_rq_pos(req) + blk_rq_sectors(req) == 1667 brq->cmd.arg = blk_rq_pos(req);
|
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 1211 sector_t sector = blk_rq_pos(rq); 1505 cmd->error = null_process_cmd(cmd, req_op(req), blk_rq_pos(req), 1556 sector_t sector = blk_rq_pos(rq);
|
/linux-master/drivers/mtd/ |
H A D | mtd_blkdevs.c | 54 block = blk_rq_pos(req) << 9 >> tr->blkshift;
|
/linux-master/drivers/cdrom/ |
H A D | gdrom.c | 583 block = blk_rq_pos(req)/GD_TO_BLK + GD_SESSION_OFFSET;
|