/linux-master/include/linux/ |
H A D | bvec.h | 22 * @bv_len: Number of bytes in the address range. 25 * The following holds for a bvec if n * PAGE_SIZE < bv_offset + bv_len: 33 unsigned int bv_len; member in struct:bio_vec 48 bv->bv_len = len; 106 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done) 117 .bv_len = mp_bvec_iter_len((bvec), (iter)), \ 136 .bv_len = bvec_iter_len((bvec), (iter)), \ 154 while (bytes && bytes >= bv[idx].bv_len) { 155 bytes -= bv[idx].bv_len; 166 * across multiple bvec entries, i.e. bytes <= bv[i->bi_idx].bv_len [all...] |
H A D | iov_iter.h | 120 (size_t)(p->bv_len - skip), 128 if (skip >= p->bv_len) {
|
H A D | bio.h | 110 /* @bytes should be less or equal to bvec[i->bi_idx].bv_len */ 131 * complete doesn't align with a bvec boundary, then bv_len and bv_offset will 149 bio_advance_iter_single((bio), &(iter), (bvl).bv_len)) 158 bio_advance_iter_single((bio), &(iter), (bvl).bv_len)) 172 #define bio_iter_last(bvec, iter) ((iter).bi_size == (bvec).bv_len) 297 fi->_seg_count = bvec->bv_len;
|
/linux-master/drivers/xen/ |
H A D | biomerge.c | 15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2;
|
/linux-master/drivers/md/ |
H A D | dm-io-rewind.c | 27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { 28 bytes -= bv[idx].bv_len; 41 iter->bi_bvec_done = bv[idx].bv_len - bytes;
|
H A D | dm-ebs-target.c | 71 unsigned int bv_len = bv->bv_len; local 76 if (unlikely(!bv->bv_page || !bv_len)) 82 while (bv_len) { 83 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); 86 if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) 113 bv_len -= cur_len;
|
H A D | dm-log-writes.c | 367 block->vecs[i].bv_len, 0); 368 if (ret != block->vecs[i].bv_len) { 380 block->vecs[i].bv_len, 0); 381 if (ret != block->vecs[i].bv_len) { 387 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; 750 block->vecs[i].bv_len = bv.bv_len;
|
/linux-master/drivers/block/ |
H A D | n64cart.c | 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); 99 pos += bvec.bv_len;
|
/linux-master/block/ |
H A D | blk-integrity.c | 40 if (seg_size + iv.bv_len > queue_max_segment_size(q)) 43 seg_size += iv.bv_len; 47 seg_size = iv.bv_len; 82 if (sg->length + iv.bv_len > queue_max_segment_size(q)) 85 sg->length += iv.bv_len; 95 sg_set_page(sg, iv.bv_page, iv.bv_len, iv.bv_offset);
|
H A D | blk-map.c | 58 bvec->bv_len, 64 if (ret < bvec->bv_len) 89 bvec->bv_len, 95 if (ret < bvec->bv_len) 366 len += bio->bi_io_vec[i].bv_len; 457 p += bvec->bv_len; 600 if (nsegs >= nr_segs || bytes > UINT_MAX - bv->bv_len) 602 if (bytes + bv->bv_len > nr_iter) 604 if (bv->bv_offset + bv->bv_len > PAGE_SIZE) 608 bytes += bv->bv_len; [all...] |
H A D | bounce.c | 91 * fromvec->bv_offset and fromvec->bv_len might have 93 * copy, bounce_copy_vec already uses tovec->bv_len 98 bio_advance_iter(from, &from_iter, tovec.bv_len); 118 bio_advance_iter(bio_orig, &orig_iter, orig_vec.bv_len); 215 bytes += from.bv_len;
|
H A D | blk-merge.c | 32 if (bv->bv_len == bio->bi_iter.bi_size) 49 bv->bv_len = iter.bi_bvec_done; 232 unsigned len = min(bv->bv_len, max_len); 252 return len > 0 || bv->bv_len > max_len; 291 bytes + bv.bv_len <= max_bytes && 292 bv.bv_offset + bv.bv_len <= PAGE_SIZE) { 294 bytes += bv.bv_len; 462 unsigned nbytes = bvec->bv_len; 497 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); 507 int nbytes = bvec->bv_len; [all...] |
H A D | t10-pi.c | 165 for (j = 0; j < iv.bv_len; j += tuple_sz) { 213 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) { 417 for (j = 0; j < iv.bv_len; j += tuple_sz) { 454 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) {
|
/linux-master/drivers/md/bcache/ |
H A D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, 251 base += bv->bv_len; 254 size -= bv->bv_len;
|
H A D | debug.c | 137 cache_set_err_on(memcmp(p1, p2, bv.bv_len), 145 bio_advance_iter(check, &citer, bv.bv_len);
|
/linux-master/fs/squashfs/ |
H A D | block.c | 50 int bytes_to_copy = min_t(int, bvec->bv_len - offset, 69 if (offset >= bvec->bv_len) { 303 if (offset < bvec->bv_len - 1) {
|
H A D | zstd_wrapper.c | 100 avail = min(length, ((int)bvec->bv_len) - offset);
|
H A D | zlib_wrapper.c | 83 avail = min(length, ((int)bvec->bv_len) - offset);
|
H A D | lzo_wrapper.c | 77 int avail = min(bytes, ((int)bvec->bv_len) - offset);
|
H A D | lz4_wrapper.c | 102 int avail = min(bytes, ((int)bvec->bv_len) - offset);
|
H A D | xz_wrapper.c | 152 avail = min(length, ((int)bvec->bv_len) - offset);
|
/linux-master/net/ceph/ |
H A D | messenger_v2.c | 156 it->bvec->bv_len - it->iov_offset), 173 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, bv.bv_len); 237 iov_iter_bvec(&con->v2.in_iter, ITER_DEST, &con->v2.in_bvec, 1, bv->bv_len); 282 con->v2.out_bvec.bv_len); 294 con->v2.out_bvec.bv_len); 908 ceph_msg_data_advance(cursor, bv.bv_len); 956 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); 959 ceph_msg_data_advance(cursor, bv.bv_len); 1115 len = min_t(int, len, bv.bv_len); 1893 con->v2.in_bvec.bv_len); [all...] |
/linux-master/fs/netfs/ |
H A D | iterator.c | 123 len = bvecs[ix].bv_len; 132 len = min3(n, bvecs[ix].bv_len - skip, max_size);
|
/linux-master/drivers/s390/block/ |
H A D | dasd_fba.c | 457 if (bv.bv_len & (blksize - 1)) 460 count += bv.bv_len >> (block->s2b_shift + 9); 461 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) 462 cidaw += bv.bv_len / blksize; 504 memcpy(copy + bv.bv_offset, dst, bv.bv_len); 508 for (off = 0; off < bv.bv_len; off += blksize) { 582 for (off = 0; off < bv.bv_len; off += blksize) { 593 memcpy(dst, cda, bv.bv_len);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io.h | 26 size_t len = min_t(size_t, folio_size(folio) - offset, bv.bv_len);
|