/linux-master/fs/btrfs/ |
H A D | extent_io.c | 291 processed_end = page_offset(page) + PAGE_SIZE - 1; 342 ASSERT(!(orig_start >= page_offset(locked_page) + PAGE_SIZE || 384 max_bytes = PAGE_SIZE; 439 start + len <= page_offset(page) + PAGE_SIZE); 577 btrfs_subpage_start_reader(fs_info, folio, page_offset(page), PAGE_SIZE); 720 eb->folio_size = PAGE_SIZE; 813 ASSERT(pg_offset + size <= PAGE_SIZE); 895 if (fs_info->nodesize >= PAGE_SIZE) { 998 const u64 end = start + PAGE_SIZE - 1; 1021 iosize = PAGE_SIZE 1485 PAGE_SIZE, !ret); local [all...] |
/linux-master/kernel/dma/ |
H A D | swiotlb.c | 331 tlb = memblock_alloc(bytes, PAGE_SIZE); 333 tlb = memblock_alloc_low(bytes, PAGE_SIZE); 399 mem->slots = memblock_alloc(alloc_size, PAGE_SIZE); 402 __func__, alloc_size, PAGE_SIZE); 494 (PAGE_SIZE << order) >> 20); 901 sz = min_t(size_t, PAGE_SIZE - offset, size); 1036 * Historically, swiotlb allocations >= PAGE_SIZE were guaranteed to be 1043 if (!alloc_align_mask && !iotlb_align_mask && alloc_size >= PAGE_SIZE) 1044 alloc_align_mask = PAGE_SIZE - 1;
|
/linux-master/fs/kernfs/ |
H A D | file.c | 224 ssize_t len = min_t(size_t, iov_iter_count(iter), PAGE_SIZE); 304 len = min_t(size_t, len, PAGE_SIZE); 672 int len = of->atomic_write_len ?: PAGE_SIZE;
|
/linux-master/fs/bcachefs/ |
H A D | sysfs.c | 68 ret = min_t(size_t, out.pos, PAGE_SIZE - 1); \
|
H A D | super-io.c | 883 bch2_bio_map(bio, ca->sb_read_scratch, PAGE_SIZE);
|
H A D | fs-io.c | 466 else if (iattr->ia_size & (PAGE_SIZE - 1)) 468 round_down(iattr->ia_size, PAGE_SIZE),
|
H A D | ec.c | 731 DIV_ROUND_UP(bytes, PAGE_SIZE));
|
H A D | checksum.c | 142 unsigned pg_len = min_t(size_t, len, PAGE_SIZE - offset); 370 unsigned page_len = min_t(unsigned, b_len, PAGE_SIZE);
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 2123 int bit = nr + (log % PAGE_SIZE) * 8;
|
/linux-master/arch/x86/kernel/cpu/ |
H A D | common.c | 2012 fixed_percpu_data) __aligned(PAGE_SIZE) __visible;
|
/linux-master/arch/x86/kernel/apic/ |
H A D | apic.c | 2646 lapic_resource.end = lapic_resource.start + PAGE_SIZE - 1;
|
/linux-master/arch/x86/events/ |
H A D | core.c | 2601 return snprintf(buf, PAGE_SIZE, "%d\n", x86_pmu_max_precise());
|
/linux-master/io_uring/ |
H A D | io_uring.c | 2733 if (uaddr & (PAGE_SIZE - 1) || !size) 2736 nr_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; 2772 page_addr += PAGE_SIZE;
|
/linux-master/fs/zonefs/ |
H A D | super.c | 1124 __bio_add_page(&bio, page, PAGE_SIZE, 0);
|
/linux-master/fs/smb/client/ |
H A D | inode.c | 56 PAGE_SIZE + MAX_CIFS_HDR_SIZE) 2719 unsigned offset = from & (PAGE_SIZE - 1); 2727 zero_user_segment(page, offset, PAGE_SIZE);
|
/linux-master/fs/ceph/ |
H A D | mds_client.h | 174 #define CEPH_CAPS_PER_RELEASE ((PAGE_SIZE - sizeof(u32) - \
|
H A D | mds_client.c | 28 #define RECONNECT_MAX_SIZE (INT_MAX - PAGE_SIZE) 2342 PAGE_SIZE, GFP_NOFS, false); 2564 num_entries = (PAGE_SIZE << order) / size; 2567 rinfo->dir_buf_size = PAGE_SIZE << order;
|
H A D | addr.c | 1156 PAGE_SIZE, 0, 1640 if (off >= PAGE_SIZE) { 1641 /* does not support inline data > PAGE_SIZE */ 1662 if (err < PAGE_SIZE) 1663 zero_user_segment(page, err, PAGE_SIZE); 1820 if (len < PAGE_SIZE) 1821 zero_user_segment(page, len, PAGE_SIZE); 2102 osd_req_op_raw_data_in_pages(rd_req, 0, pages, PAGE_SIZE,
|
/linux-master/drivers/pci/ |
H A D | quirks.c | 575 if (r->flags & IORESOURCE_MEM && resource_size(r) < PAGE_SIZE) { 576 r->end = PAGE_SIZE - 1;
|
/linux-master/drivers/md/ |
H A D | raid1.c | 1217 unsigned vcnt = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; 1232 int len = min_t(int, PAGE_SIZE, size); 1583 BIO_MAX_VECS * (PAGE_SIZE >> 9)); 2147 if (s > (PAGE_SIZE>>9)) 2148 s = PAGE_SIZE >> 9; 2254 vcnt = (r1_bio->sectors + PAGE_SIZE / 512 - 1) >> (PAGE_SHIFT - 9); 2396 if (s > (PAGE_SIZE>>9)) 2397 s = PAGE_SIZE >> 9; 2983 int len = PAGE_SIZE;
|
/linux-master/drivers/iommu/intel/ |
H A D | svm.c | 252 (end - start + PAGE_SIZE - 1) >> VTD_PAGE_SHIFT, 0);
|
H A D | iommu.c | 2495 old_ce = memremap(old_ce_phys, PAGE_SIZE, 2563 old_rt = memremap(old_rt_phys, PAGE_SIZE, MEMREMAP_WB); 2607 __iommu_flush_cache(iommu, iommu->root_entry, PAGE_SIZE); 2928 if (!IS_ALIGNED(rmrr->base_address, PAGE_SIZE) || 2929 !IS_ALIGNED(rmrr->end_address + 1, PAGE_SIZE) || 3839 domain_flush_cache(domain, domain->pgd, PAGE_SIZE); 4047 /* Round up size to next multiple of PAGE_SIZE, if it and
|
/linux-master/drivers/iommu/amd/ |
H A D | iommu.c | 1094 pages = iommu_num_pages(address, size, PAGE_SIZE);
|
H A D | init.c | 3852 for (page = va; page < (va + size); page += PAGE_SIZE) { 3878 ret = iommu_make_shared((void *)iommu->cmd_sem, PAGE_SIZE);
|
/linux-master/drivers/ata/ |
H A D | libata-scsi.c | 1072 * PAGE_SIZE. ATAPI devices also need the alignment as 1075 if (sdev->sector_size > PAGE_SIZE) 1077 "sector_size=%u > PAGE_SIZE, PIO may malfunction\n",
|