/linux-master/include/linux/ |
H A D | bvec.h | 39 * @bv: bvec to initialize 44 static inline void bvec_set_page(struct bio_vec *bv, struct page *page, argument 47 bv->bv_page = page; 48 bv->bv_len = len; 49 bv->bv_offset = offset; 54 * @bv: bvec to initialize 59 static inline void bvec_set_folio(struct bio_vec *bv, struct folio *folio, argument 62 bvec_set_page(bv, &folio->page, len, offset); 67 * @bv: bvec to initialize 71 static inline void bvec_set_virt(struct bio_vec *bv, voi argument 89 struct bio_vec bv; member in struct:bvec_iter_all 140 bvec_iter_advance(const struct bio_vec *bv, struct bvec_iter *iter, unsigned bytes) argument 168 bvec_iter_advance_single(const struct bio_vec *bv, struct bvec_iter *iter, unsigned int bytes) argument 207 struct bio_vec *bv = &iter_all->bv; local [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | gen7_renderclear.c | 44 static int num_primitives(const struct batch_vals *bv) argument 52 return bv->max_threads; 56 batch_get_defaults(struct drm_i915_private *i915, struct batch_vals *bv) argument 62 bv->max_threads = 70; 65 bv->max_threads = 140; 68 bv->max_threads = 280; 71 bv->surface_height = 16 * 16; 72 bv->surface_width = 32 * 2 * 16; 77 bv->max_threads = 36; 80 bv 142 gen7_fill_surface_state(struct batch_chunk *state, const u32 dst_offset, const struct batch_vals *bv) argument 175 gen7_fill_binding_table(struct batch_chunk *state, const struct batch_vals *bv) argument 207 gen7_fill_interface_descriptor(struct batch_chunk *state, const struct batch_vals *bv, const struct cb_kernel *kernel, unsigned int count) argument 261 gen7_emit_vfe_state(struct batch_chunk *batch, const struct batch_vals *bv, u32 urb_size, u32 curbe_size, u32 mode) argument 376 emit_batch(struct i915_vma * const vma, u32 *start, const struct batch_vals *bv) argument 434 struct batch_vals bv; local [all...] |
/linux-master/drivers/md/bcache/ |
H A D | util.c | 234 struct bio_vec *bv = bio->bi_io_vec; local 239 bv->bv_offset = base ? offset_in_page(base) : 0; 242 for (; size; bio->bi_vcnt++, bv++) { 243 bv->bv_offset = 0; 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, 247 bv->bv_page = is_vmalloc_addr(base) 251 base += bv->bv_len; 254 size -= bv->bv_len; 271 struct bio_vec *bv; local [all...] |
H A D | debug.c | 112 struct bio_vec bv, cbv; local 130 bio_for_each_segment(bv, bio, iter) { 131 void *p1 = bvec_kmap_local(&bv); 137 cache_set_err_on(memcmp(p1, p2, bv.bv_len), 145 bio_advance_iter(check, &citer, bv.bv_len);
|
/linux-master/drivers/block/ |
H A D | n64cart.c | 62 static bool n64cart_do_bvec(struct device *dev, struct bio_vec *bv, u32 pos) argument 68 WARN_ON_ONCE((bv->bv_offset & (MIN_ALIGNMENT - 1)) || 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); 71 dma_addr = dma_map_bvec(dev, bv, DMA_FROM_DEVICE, 0); 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE);
|
/linux-master/drivers/md/ |
H A D | dm-io-rewind.c | 12 static inline bool dm_bvec_iter_rewind(const struct bio_vec *bv, argument 27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { 28 bytes -= bv[idx].bv_len; 41 iter->bi_bvec_done = bv[idx].bv_len - bytes;
|
H A D | dm-ebs-target.c | 65 static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv, argument 71 unsigned int bv_len = bv->bv_len; 76 if (unlikely(!bv->bv_page || !bv_len)) 79 pa = bvec_virt(bv); 102 flush_dcache_page(bv->bv_page); 104 flush_dcache_page(bv->bv_page); 125 struct bio_vec bv; local 128 bio_for_each_bvec(bv, bio, iter) { 129 rr = __ebs_rw_bvec(ec, op, &bv, &iter);
|
/linux-master/arch/parisc/kernel/ |
H A D | real2.S | 79 bv 0(%r31) 101 bv 0(%rp) 121 bv 0(%r2) 136 bv 0(%r2) 177 bv 0(%r2) 214 bv 0(%r2) 260 bv 0(%r31) 273 bv 0(%rp) 288 bv %r0(%r2)
|
H A D | entry.S | 161 bv,n 0(%r3) 772 bv %r0(%r2) 781 bv %r0(%r2) 940 bv %r0(%r20) 969 bv %r0(%r1) 1701 bv %r0(%r2) 1897 bv %r0(%r19) /* jumps to schedule() */ 1928 bv %r0(%rp) 1996 bv,n (%r1) 2092 bv, [all...] |
H A D | hpmc.S | 154 bv (r3) /* call pdce_proc */ 169 bv (%r3) /* call pdce_proc */ 199 bv (%r3) /* call pdce_proc */ 221 bv (%r5) 269 bv (%r3) /* call pdce_proc */
|
H A D | head.S | 101 bv,n (%r1) 208 bv,n (%rp) 250 bv (%rp) 263 bv (%r3) 277 bv (%r3) 396 bv (%rp)
|
/linux-master/fs/bcachefs/ |
H A D | fs-io.h | 20 static inline struct folio_vec biovec_to_foliovec(struct bio_vec bv) argument 23 struct folio *folio = page_folio(bv.bv_page); 24 size_t offset = (folio_page_idx(folio, bv.bv_page) << PAGE_SHIFT) + 25 bv.bv_offset; 26 size_t len = min_t(size_t, folio_size(folio) - offset, bv.bv_len);
|
H A D | compress.c | 43 struct bio_vec bv; local 47 __bio_for_each_bvec(bv, bio, iter, start) { 49 expected_start != page_address(bv.bv_page) + bv.bv_offset) 52 expected_start = page_address(bv.bv_page) + 53 bv.bv_offset + bv.bv_len; 63 struct bio_vec bv; local 81 __bio_for_each_segment(bv, bio, iter, start) { 83 bv [all...] |
H A D | checksum.c | 252 struct bio_vec bv; local 268 __bio_for_each_segment(bv, bio, *iter, *iter) { 269 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; 271 bch2_checksum_update(&state, p, bv.bv_len); 275 __bio_for_each_bvec(bv, bio, *iter, *iter) 276 bch2_checksum_update(&state, page_address(bv.bv_page) + bv.bv_offset, 277 bv.bv_len); 291 __bio_for_each_segment(bv, bi 324 struct bio_vec bv; local [all...] |
/linux-master/block/ |
H A D | blk-merge.c | 21 static inline void bio_get_first_bvec(struct bio *bio, struct bio_vec *bv) argument 23 *bv = mp_bvec_iter_bvec(bio->bi_io_vec, bio->bi_iter); 26 static inline void bio_get_last_bvec(struct bio *bio, struct bio_vec *bv) argument 31 bio_get_first_bvec(bio, bv); 32 if (bv->bv_len == bio->bi_iter.bi_size) 42 *bv = bio->bi_io_vec[idx]; 49 bv->bv_len = iter.bi_bvec_done; 210 * @bv: [in] bvec to examine 212 * by the number of segments from @bv that may be appended to that 215 * by the number of bytes from @bv tha 227 bvec_split_segs(const struct queue_limits *lim, const struct bio_vec *bv, unsigned *nsegs, unsigned *bytes, unsigned max_segs, unsigned max_bytes) argument 278 struct bio_vec bv, bvprv, *bvprvp = NULL; local 413 struct bio_vec bv; local 493 __blk_bvec_map_sg(struct bio_vec bv, struct scatterlist *sglist, struct scatterlist **sg) argument [all...] |
H A D | bio.c | 162 void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned short nr_vecs) argument 167 mempool_free(bv, pool); 169 kmem_cache_free(biovec_slab(nr_vecs)->slab, bv); 611 struct bio_vec bv; local 614 __bio_for_each_segment(bv, bio, iter, start) 615 memzero_bvec(&bv); 631 struct bio_vec bv; local 642 bio_for_each_segment(bv, bio, iter) { 643 if (done + bv.bv_len > new_size) { 650 zero_user(bv 909 bvec_try_merge_page(struct bio_vec *bv, struct page *page, unsigned int len, unsigned int off, bool *same_page) argument 940 bvec_try_merge_hw_page(struct request_queue *q, struct bio_vec *bv, struct page *page, unsigned len, unsigned offset, bool *same_page) argument 982 struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt - 1]; local 1246 struct bio_vec *bv = bio->bi_io_vec + bio->bi_vcnt; local [all...] |
H A D | bio-integrity.c | 94 static void bio_integrity_unpin_bvec(struct bio_vec *bv, int nr_vecs, argument 100 if (dirty && !PageCompound(bv[i].bv_page)) 101 set_page_dirty_lock(bv[i].bv_page); 102 unpin_user_page(bv[i].bv_page); 177 struct bio_vec *bv = &bip->bip_vec[bip->bip_vcnt - 1]; local 180 if (bvec_try_merge_hw_page(q, bv, page, len, offset, 194 if (bvec_gap_to_prev(&q->limits, bv, offset)) 389 struct bio_vec bv; local 400 __bio_for_each_segment(bv, bio, bviter, *proc_iter) { 401 void *kaddr = bvec_kmap_local(&bv); [all...] |
H A D | blk-crypto-fallback.c | 164 struct bio_vec bv; local 179 bio_for_each_segment(bv, bio_src, iter) 180 bio->bi_io_vec[bio->bi_vcnt++] = bv; 216 struct bio_vec bv; local 219 bio_for_each_segment(bv, bio, iter) { 220 num_sectors += bv.bv_len >> SECTOR_SHIFT; 389 struct bio_vec bv; local 418 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { 419 struct page *page = bv.bv_page; 421 sg_set_page(&sg, page, data_unit_size, bv [all...] |
H A D | blk-crypto.c | 218 struct bio_vec bv; local 220 bio_for_each_segment(bv, bio, iter) { 221 if (!IS_ALIGNED(bv.bv_len | bv.bv_offset, data_unit_size))
|
H A D | blk-map.c | 540 struct bio_vec bv; local 543 bio_for_each_bvec(bv, bio, iter) 589 struct bio_vec *bv = &bvecs[i]; local 595 if (bvprvp && bvec_gap_to_prev(lim, bvprvp, bv->bv_offset)) { 600 if (nsegs >= nr_segs || bytes > UINT_MAX - bv->bv_len) 602 if (bytes + bv->bv_len > nr_iter) 604 if (bv->bv_offset + bv->bv_len > PAGE_SIZE) 608 bytes += bv->bv_len; 609 bvprvp = bv; [all...] |
/linux-master/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | main.c | 66 struct nfp_bpf_vnic *bv; local 79 bv = kzalloc(sizeof(*bv), GFP_KERNEL); 80 if (!bv) 82 nn->app_priv = bv; 88 bv->start_off = nn_readw(nn, NFP_NET_CFG_BPF_START); 89 bv->tgt_done = nn_readw(nn, NFP_NET_CFG_BPF_DONE); 99 struct nfp_bpf_vnic *bv = nn->app_priv; local 101 WARN_ON(bv->tc_prog); 102 kfree(bv); 111 struct nfp_bpf_vnic *bv; local 185 struct nfp_bpf_vnic *bv; local [all...] |
/linux-master/fs/netfs/ |
H A D | iterator.c | 40 struct bio_vec *bv = NULL; local 54 bv_size = array_size(max_pages, sizeof(*bv)); 55 bv = kvmalloc(bv_size, GFP_KERNEL); 56 if (!bv) 64 pages = (void *)bv + bv_size - pg_size; 92 bvec_set_page(bv + npages + i, *pages++, len - offset, offset); 100 iov_iter_bvec(new, orig->data_source, bv, npages, orig_len - count);
|
/linux-master/arch/parisc/boot/compressed/ |
H A D | head.S | 71 bv,n 0(%r3) 84 bv,n 0(%ret0)
|
/linux-master/drivers/s390/block/ |
H A D | dasd_fba.c | 435 struct bio_vec bv; local 456 rq_for_each_segment(bv, req, iter) { 457 if (bv.bv_len & (blksize - 1)) 460 count += bv.bv_len >> (block->s2b_shift + 9); 461 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) 462 cidaw += bv.bv_len / blksize; 498 rq_for_each_segment(bv, req, iter) { 499 dst = bvec_virt(&bv); 504 memcpy(copy + bv 567 struct bio_vec bv; local [all...] |
/linux-master/net/ceph/ |
H A D | messenger_v2.c | 146 struct bio_vec bv; local 154 bvec_set_page(&bv, it->bvec->bv_page, 168 if (sendpage_ok(bv.bv_page)) 173 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, bv.bv_len); 232 static void set_in_bvec(struct ceph_connection *con, const struct bio_vec *bv) argument 236 con->v2.in_bvec = *bv; 237 iov_iter_bvec(&con->v2.in_iter, ITER_DEST, &con->v2.in_bvec, 1, bv->bv_len); 273 static void set_out_bvec(struct ceph_connection *con, const struct bio_vec *bv, argument 279 con->v2.out_bvec = *bv; 859 get_bvec_at(struct ceph_msg_data_cursor *cursor, struct bio_vec *bv) argument 897 struct bio_vec bv; local 949 struct bio_vec bv; local 1112 struct bio_vec bv; local 1862 struct bio_vec bv; local 1888 struct bio_vec bv; local 1929 struct bio_vec bv; local 2079 struct bio_vec bv; local 3189 struct bio_vec bv; local 3202 struct bio_vec bv; local 3227 struct bio_vec bv; local [all...] |