Searched refs:blk_rq_pos (Results 1 - 25 of 46) sorted by relevance

12

/linux-master/include/linux/
H A Dt10-pi.h48 return blk_rq_pos(rq) >> (shift - SECTOR_SHIFT) & 0xffffffff;
79 return lower_48_bits(blk_rq_pos(rq) >> (shift - SECTOR_SHIFT));
H A Dblktrace_api.h121 if (blk_rq_is_passthrough(rq) || blk_rq_pos(rq) == (sector_t)-1)
123 return blk_rq_pos(rq);
H A Dblk-mq.h1029 * blk_rq_pos() : the current sector
1036 static inline sector_t blk_rq_pos(const struct request *rq) function
1156 return disk_zone_no(rq->q->disk, blk_rq_pos(rq));
1161 return disk_zone_is_seq(rq->q->disk, blk_rq_pos(rq));
/linux-master/block/
H A Delevator.c54 #define rq_hash_key(rq) (blk_rq_pos(rq) + blk_rq_sectors(rq))
237 if (blk_rq_pos(rq) < blk_rq_pos(__rq))
239 else if (blk_rq_pos(rq) >= blk_rq_pos(__rq))
264 if (sector < blk_rq_pos(rq))
266 else if (sector > blk_rq_pos(rq))
357 __rq = elv_rqhash_find(q, blk_rq_pos(rq));
H A Dblk-merge.c642 blk_rq_get_max_sectors(req, blk_rq_pos(req))) {
677 blk_rq_get_max_sectors(req, blk_rq_pos(req)))
699 blk_rq_get_max_sectors(req, blk_rq_pos(req)))
791 else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next))
958 else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector)
960 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector)
1043 blk_rq_get_max_sectors(req, blk_rq_pos(req)))
H A Dmq-deadline.c160 * Return the first request for which blk_rq_pos() >= @pos. For zoned devices,
183 if (blk_rq_pos(rq) >= pos) {
321 return blk_rq_pos(prev) + blk_rq_sectors(prev) == blk_rq_pos(rq);
331 sector_t pos = blk_rq_pos(rq);
336 } while (rq && blk_rq_pos(rq) == pos);
370 rb_rq = deadline_from_pos(per_prio, data_dir, blk_rq_pos(rq));
371 if (rb_rq && blk_rq_pos(rb_rq) < blk_rq_pos(rq))
550 dd->per_prio[prio].latest_pos[data_dir] = blk_rq_pos(r
[all...]
H A Dbfq-iosched.c518 s1 = blk_rq_pos(rq1);
519 s2 = blk_rq_pos(rq2);
747 if (sector > blk_rq_pos(bfqq->next_rq))
749 else if (sector < blk_rq_pos(bfqq->next_rq))
813 blk_rq_pos(bfqq->next_rq), &parent, &p);
1039 return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last));
2375 return abs(blk_rq_pos(rq) - last_pos);
2508 blk_rq_pos(req) <
2509 blk_rq_pos(container_of(rb_prev(&req->rb_node),
2669 return blk_rq_pos(io_struc
[all...]
/linux-master/include/scsi/
H A Dscsi_cmnd.h225 return blk_rq_pos(scsi_cmd_to_rq(scmd));
232 return blk_rq_pos(scsi_cmd_to_rq(scmd)) >> shift;
/linux-master/drivers/block/
H A Dz2ram.c72 unsigned long start = blk_rq_pos(req) << 9;
80 (unsigned long long)blk_rq_pos(req),
H A Dvirtio_blk.c185 range[0].sector = cpu_to_le64(blk_rq_pos(req));
257 sector = blk_rq_pos(req);
261 sector = blk_rq_pos(req);
278 sector = blk_rq_pos(req);
282 sector = blk_rq_pos(req);
286 sector = blk_rq_pos(req);
290 sector = blk_rq_pos(req);
295 sector = blk_rq_pos(req);
H A Dswim3.c343 * Do not remove the cast. blk_rq_pos(req) is now a sector_t and can be
347 fs->req_cyl = ((long)blk_rq_pos(req)) / fs->secpercyl;
348 x = ((long)blk_rq_pos(req)) % fs->secpercyl;
649 (long)blk_rq_pos(fs->cur_req));
767 (long)blk_rq_pos(req), err);
H A Dps3disk.c116 start_sector = blk_rq_pos(req) * priv->blocking_factor;
/linux-master/drivers/md/
H A Ddm-rq.c134 blk_rq_pos(orig), tio->n_sectors, true,
385 blk_rq_pos(rq));
443 blk_rq_pos(orig), tio->n_sectors, false, 0,
/linux-master/drivers/s390/block/
H A Ddasd_fba.c333 first_rec = blk_rq_pos(req) >> block->s2b_shift;
335 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
450 first_rec = blk_rq_pos(req) >> block->s2b_shift;
452 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
488 block->bp_block, blk_rq_pos(req), blk_rq_sectors(req));
H A Ddasd_diag.c526 first_rec = blk_rq_pos(req) >> block->s2b_shift;
528 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
H A Dscm_blk.c196 msb->scm_addr = scmdev->address + ((u64) blk_rq_pos(req) << 9);
/linux-master/include/trace/events/
H A Dblock.h122 __entry->sector = blk_rq_pos(rq);
567 __entry->sector = blk_rq_pos(rq);
/linux-master/drivers/scsi/
H A Dsd_zbc.c334 sector_t sector = blk_rq_pos(rq);
481 sector_t sector = blk_rq_pos(rq);
H A Dsr.c332 good_bytes = (error_sector - blk_rq_pos(rq)) << 9;
431 if (((unsigned int)blk_rq_pos(rq) % (s_size >> 9)) ||
447 block = (unsigned int)blk_rq_pos(rq) / (s_size >> 9);
H A Dsd.c895 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq));
926 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq));
953 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq));
979 u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq));
1216 sector_t lba = sectors_to_logical(sdp, blk_rq_pos(rq));
1237 if (blk_rq_pos(rq) + blk_rq_sectors(rq) > get_capacity(rq->q->disk)) {
1242 if ((blk_rq_pos(rq) & mask) || (blk_rq_sectors(rq) & mask)) {
1310 (unsigned long long)blk_rq_pos(rq),
2144 start_lba = sectors_to_logical(sdev, blk_rq_pos(req));
/linux-master/drivers/nvme/host/
H A Dzns.c244 c->zms.slba = cpu_to_le64(nvme_sect_to_lba(ns->head, blk_rq_pos(req)));
/linux-master/drivers/mmc/core/
H A Dblock.c1165 from = blk_rq_pos(req);
1220 from = blk_rq_pos(req);
1299 if (!IS_ALIGNED(blk_rq_pos(req), card->ext_csd.rel_sectors))
1401 brq->data.blk_addr = blk_rq_pos(req);
1425 (blk_rq_pos(req) + blk_rq_sectors(req) ==
1667 brq->cmd.arg = blk_rq_pos(req);
/linux-master/drivers/block/null_blk/
H A Dmain.c1211 sector_t sector = blk_rq_pos(rq);
1505 cmd->error = null_process_cmd(cmd, req_op(req), blk_rq_pos(req),
1556 sector_t sector = blk_rq_pos(rq);
/linux-master/drivers/mtd/
H A Dmtd_blkdevs.c54 block = blk_rq_pos(req) << 9 >> tr->blkshift;
/linux-master/drivers/cdrom/
H A Dgdrom.c583 block = blk_rq_pos(req)/GD_TO_BLK + GD_SESSION_OFFSET;

Completed in 265 milliseconds

12