/linux-master/include/linux/ |
H A D | bvec.h | 21 * @bv_page: First page associated with the address range. 23 * @bv_offset: Start of the address range relative to the start of @bv_page. 27 * nth_page(@bv_page, n) == @bv_page + n 32 struct page *bv_page; member in struct:bio_vec 47 bv->bv_page = page; 102 (__bvec_iter_bvec((bvec), (iter))->bv_page) 116 .bv_page = mp_bvec_iter_page((bvec), (iter)), \ 135 .bv_page = bvec_iter_page((bvec), (iter)), \ 210 bv->bv_page [all...] |
H A D | bio.h | 253 return bio_first_bvec_all(bio)->bv_page; 294 fi->folio = page_folio(bvec->bv_page); 296 PAGE_SIZE * (bvec->bv_page - &fi->folio->page);
|
H A D | iov_iter.h | 117 void *kaddr = kmap_local_page(p->bv_page + offset / PAGE_SIZE);
|
/linux-master/drivers/xen/ |
H A D | biomerge.c | 12 unsigned long bfn1 = pfn_to_bfn(page_to_pfn(vec1->bv_page));
|
/linux-master/block/ |
H A D | bounce.c | 89 if (tovec.bv_page != fromvec.bv_page) { 95 memcpy_to_bvec(&tovec, page_address(fromvec.bv_page) + 114 if (bvec->bv_page != orig_vec.bv_page) { 115 dec_zone_page_state(bvec->bv_page, NR_BOUNCE); 116 mempool_free(bvec->bv_page, &page_pool); 216 if (PageHighMem(from.bv_page)) 245 if (!PageHighMem(to->bv_page)) 252 flush_dcache_page(to->bv_page); [all...] |
H A D | blk-crypto-fallback.c | 150 mempool_free(enc_bio->bi_io_vec[i].bv_page, 319 struct page *plaintext_page = enc_bvec->bv_page; 323 enc_bvec->bv_page = ciphertext_page; 360 mempool_free(enc_bio->bi_io_vec[--i].bv_page, 419 struct page *page = bv.bv_page;
|
H A D | bio-integrity.c | 100 if (dirty && !PageCompound(bv[i].bv_page)) 101 set_page_dirty_lock(bv[i].bv_page); 102 unpin_user_page(bv[i].bv_page);
|
H A D | blk.h | 100 phys_addr_t addr1 = page_to_phys(vec1->bv_page) + vec1->bv_offset; 101 phys_addr_t addr2 = page_to_phys(vec2->bv_page) + vec2->bv_offset; 112 if (xen_domain() && !xen_biovec_phys_mergeable(vec1, vec2->bv_page))
|
H A D | bio.c | 650 zero_user(bv.bv_page, bv.bv_offset + offset, 913 phys_addr_t vec_end_addr = page_to_phys(bv->bv_page) + bv_end - 1; 920 if (!zone_device_pages_have_same_pgmap(bv->bv_page, page)) 927 if (bv->bv_page + bv_end / PAGE_SIZE != page + off / PAGE_SIZE) 945 phys_addr_t addr1 = page_to_phys(bv->bv_page) + bv->bv_offset; 1139 * after the bv_page. BIOs do not support folios that are 4GiB or larger. 1441 __free_page(bvec->bv_page);
|
H A D | blk-merge.c | 237 seg_size = get_max_segment_size(lim, bv->bv_page, 468 bvec->bv_page, offset), nbytes); 469 struct page *page = bvec->bv_page; 497 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset);
|
H A D | blk-map.c | 56 ret = copy_page_from_iter(bvec->bv_page, 87 ret = copy_page_to_iter(bvec->bv_page,
|
/linux-master/drivers/md/bcache/ |
H A D | util.c | 247 bv->bv_page = is_vmalloc_addr(base) 278 bv->bv_page = alloc_page(gfp_mask); 279 if (!bv->bv_page) { 281 __free_page(bv->bv_page);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io.h | 23 struct folio *folio = page_folio(bv.bv_page); 24 size_t offset = (folio_page_idx(folio, bv.bv_page) << PAGE_SHIFT) +
|
H A D | checksum.c | 269 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; 276 bch2_checksum_update(&state, page_address(bv.bv_page) + bv.bv_offset, 292 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; 300 page_address(bv.bv_page) + bv.bv_offset, 350 sg_set_page(sg++, bv.bv_page, bv.bv_len, bv.bv_offset);
|
H A D | compress.c | 49 expected_start != page_address(bv.bv_page) + bv.bv_offset) 52 expected_start = page_address(bv.bv_page) + 103 pages[nr_pages++] = bv.bv_page;
|
/linux-master/fs/netfs/ |
H A D | objects.c | 102 if (rreq->direct_bv[i].bv_page) { 104 unpin_user_page(rreq->direct_bv[i].bv_page);
|
H A D | io.c | 356 flush_dcache_page(rreq->direct_bv[i].bv_page); 367 flush_dcache_page(rreq->direct_bv[i].bv_page);
|
/linux-master/mm/ |
H A D | page_io.c | 267 struct page *page = sio->bvec[0].bv_page; 284 page = sio->bvec[p].bv_page; 291 end_page_writeback(sio->bvec[p].bv_page); 411 struct folio *folio = page_folio(sio->bvec[p].bv_page); 419 struct folio *folio = page_folio(sio->bvec[p].bv_page);
|
/linux-master/drivers/md/ |
H A D | dm-ebs-target.c | 76 if (unlikely(!bv->bv_page || !bv_len)) 102 flush_dcache_page(bv->bv_page); 104 flush_dcache_page(bv->bv_page);
|
H A D | dm-log-writes.c | 204 if (block->vecs[i].bv_page) 205 __free_page(block->vecs[i].bv_page); 366 ret = bio_add_page(bio, block->vecs[i].bv_page, 379 ret = bio_add_page(bio, block->vecs[i].bv_page, 749 block->vecs[i].bv_page = page;
|
/linux-master/arch/m68k/emu/ |
H A D | nfblock.c | 74 page_to_phys(bvec.bv_page) + bvec.bv_offset);
|
/linux-master/fs/erofs/ |
H A D | zdata.c | 1435 bvec->bv_page = &zbv.folio->page; 1436 DBG_BUGON(z_erofs_is_shortlived_page(bvec->bv_page)); 1475 bvec->bv_page = NULL; 1499 bvec->bv_page = page; 1642 if (!bvec.bv_page) 1660 if (unlikely(PageWorkingset(bvec.bv_page)) && 1682 if (!bio_add_page(bio, bvec.bv_page, bvec.bv_len,
|
/linux-master/net/ceph/ |
H A D | messenger_v2.c | 154 bvec_set_page(&bv, it->bvec->bv_page, 168 if (sendpage_ok(bv.bv_page)) 956 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); 1116 memcpy_page(bv.bv_page, bv.bv_offset, 1878 bv.bv_page = con->bounce_page; 1896 memcpy_to_page(bv.bv_page, bv.bv_offset, 1901 con->v2.in_bvec.bv_page, 1910 bv.bv_page = con->bounce_page; 1941 memcpy_to_page(bv.bv_page, bv.bv_offset, 1946 con->v2.in_bvec.bv_page, [all...] |
/linux-master/drivers/mtd/ |
H A D | mtd_blkdevs.c | 77 flush_dcache_page(bvec.bv_page); 84 flush_dcache_page(bvec.bv_page);
|
/linux-master/fs/squashfs/ |
H A D | block.c | 93 struct page *page = bv->bv_page;
|